var/home/core/zuul-output/0000755000175000017500000000000015071667307014541 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015071677774015516 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004665462715071677765017746 0ustar rootrootOct 09 08:17:20 crc systemd[1]: Starting Kubernetes Kubelet... Oct 09 08:17:20 crc restorecon[4660]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:20 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 09 08:17:21 crc restorecon[4660]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 09 08:17:22 crc kubenswrapper[4872]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.207332 4872 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214838 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214856 4872 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214861 4872 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214866 4872 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214870 4872 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214877 4872 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214881 4872 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214884 4872 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214888 4872 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214892 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214896 4872 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214899 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214903 4872 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214908 4872 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214913 4872 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214918 4872 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214923 4872 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214927 4872 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214932 4872 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214938 4872 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214943 4872 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214948 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214952 4872 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214955 4872 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214959 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214962 4872 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214966 4872 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214970 4872 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214974 4872 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214980 4872 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214983 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214987 4872 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214990 4872 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214995 4872 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.214998 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215002 4872 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215005 4872 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215009 4872 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215013 4872 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215016 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215020 4872 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215024 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215027 4872 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215031 4872 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215035 4872 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215038 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215042 4872 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215045 4872 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215049 4872 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215052 4872 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215056 4872 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215059 4872 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215063 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215067 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215071 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215075 4872 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215078 4872 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215081 4872 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215085 4872 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215088 4872 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215091 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215095 4872 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215099 4872 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215103 4872 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215107 4872 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215112 4872 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215116 4872 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215120 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215124 4872 feature_gate.go:330] unrecognized feature gate: Example Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215127 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.215131 4872 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215212 4872 flags.go:64] FLAG: --address="0.0.0.0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215222 4872 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215230 4872 flags.go:64] FLAG: --anonymous-auth="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215237 4872 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215244 4872 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215249 4872 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215257 4872 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215266 4872 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215271 4872 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215276 4872 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215283 4872 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215287 4872 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215292 4872 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215296 4872 flags.go:64] FLAG: --cgroup-root="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215300 4872 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215304 4872 flags.go:64] FLAG: --client-ca-file="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215308 4872 flags.go:64] FLAG: --cloud-config="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215312 4872 flags.go:64] FLAG: --cloud-provider="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215316 4872 flags.go:64] FLAG: --cluster-dns="[]" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215322 4872 flags.go:64] FLAG: --cluster-domain="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215327 4872 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215332 4872 flags.go:64] FLAG: --config-dir="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215337 4872 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215343 4872 flags.go:64] FLAG: --container-log-max-files="5" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215350 4872 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215355 4872 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215360 4872 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215366 4872 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215371 4872 flags.go:64] FLAG: --contention-profiling="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215375 4872 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215379 4872 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215384 4872 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215388 4872 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215393 4872 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215397 4872 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215401 4872 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215405 4872 flags.go:64] FLAG: --enable-load-reader="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215410 4872 flags.go:64] FLAG: --enable-server="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215415 4872 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215420 4872 flags.go:64] FLAG: --event-burst="100" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215425 4872 flags.go:64] FLAG: --event-qps="50" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215432 4872 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215438 4872 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215444 4872 flags.go:64] FLAG: --eviction-hard="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215451 4872 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215456 4872 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215461 4872 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215466 4872 flags.go:64] FLAG: --eviction-soft="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215471 4872 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215476 4872 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215480 4872 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215484 4872 flags.go:64] FLAG: --experimental-mounter-path="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215488 4872 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215492 4872 flags.go:64] FLAG: --fail-swap-on="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215496 4872 flags.go:64] FLAG: --feature-gates="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215502 4872 flags.go:64] FLAG: --file-check-frequency="20s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215506 4872 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215510 4872 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215515 4872 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215520 4872 flags.go:64] FLAG: --healthz-port="10248" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215525 4872 flags.go:64] FLAG: --help="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215530 4872 flags.go:64] FLAG: --hostname-override="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215535 4872 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215540 4872 flags.go:64] FLAG: --http-check-frequency="20s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215545 4872 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215550 4872 flags.go:64] FLAG: --image-credential-provider-config="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215557 4872 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215562 4872 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215567 4872 flags.go:64] FLAG: --image-service-endpoint="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215572 4872 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215577 4872 flags.go:64] FLAG: --kube-api-burst="100" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215582 4872 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215587 4872 flags.go:64] FLAG: --kube-api-qps="50" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215592 4872 flags.go:64] FLAG: --kube-reserved="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215597 4872 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215601 4872 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215605 4872 flags.go:64] FLAG: --kubelet-cgroups="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215610 4872 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215615 4872 flags.go:64] FLAG: --lock-file="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215620 4872 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215626 4872 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215632 4872 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215665 4872 flags.go:64] FLAG: --log-json-split-stream="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215670 4872 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215676 4872 flags.go:64] FLAG: --log-text-split-stream="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215681 4872 flags.go:64] FLAG: --logging-format="text" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215686 4872 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215692 4872 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215697 4872 flags.go:64] FLAG: --manifest-url="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215702 4872 flags.go:64] FLAG: --manifest-url-header="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215708 4872 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215714 4872 flags.go:64] FLAG: --max-open-files="1000000" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215721 4872 flags.go:64] FLAG: --max-pods="110" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215726 4872 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215731 4872 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215736 4872 flags.go:64] FLAG: --memory-manager-policy="None" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215741 4872 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215747 4872 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215753 4872 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215758 4872 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215769 4872 flags.go:64] FLAG: --node-status-max-images="50" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215775 4872 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215780 4872 flags.go:64] FLAG: --oom-score-adj="-999" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215785 4872 flags.go:64] FLAG: --pod-cidr="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215790 4872 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215798 4872 flags.go:64] FLAG: --pod-manifest-path="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215803 4872 flags.go:64] FLAG: --pod-max-pids="-1" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215809 4872 flags.go:64] FLAG: --pods-per-core="0" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215814 4872 flags.go:64] FLAG: --port="10250" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215820 4872 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215825 4872 flags.go:64] FLAG: --provider-id="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215830 4872 flags.go:64] FLAG: --qos-reserved="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215836 4872 flags.go:64] FLAG: --read-only-port="10255" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215841 4872 flags.go:64] FLAG: --register-node="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215846 4872 flags.go:64] FLAG: --register-schedulable="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215851 4872 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215859 4872 flags.go:64] FLAG: --registry-burst="10" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215865 4872 flags.go:64] FLAG: --registry-qps="5" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215870 4872 flags.go:64] FLAG: --reserved-cpus="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215875 4872 flags.go:64] FLAG: --reserved-memory="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215881 4872 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215886 4872 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215892 4872 flags.go:64] FLAG: --rotate-certificates="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215897 4872 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215902 4872 flags.go:64] FLAG: --runonce="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215907 4872 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215912 4872 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215917 4872 flags.go:64] FLAG: --seccomp-default="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215922 4872 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215927 4872 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215933 4872 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215938 4872 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215943 4872 flags.go:64] FLAG: --storage-driver-password="root" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215949 4872 flags.go:64] FLAG: --storage-driver-secure="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215954 4872 flags.go:64] FLAG: --storage-driver-table="stats" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215960 4872 flags.go:64] FLAG: --storage-driver-user="root" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215965 4872 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215971 4872 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215976 4872 flags.go:64] FLAG: --system-cgroups="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215986 4872 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.215995 4872 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216001 4872 flags.go:64] FLAG: --tls-cert-file="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216007 4872 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216013 4872 flags.go:64] FLAG: --tls-min-version="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216019 4872 flags.go:64] FLAG: --tls-private-key-file="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216024 4872 flags.go:64] FLAG: --topology-manager-policy="none" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216030 4872 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216035 4872 flags.go:64] FLAG: --topology-manager-scope="container" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216040 4872 flags.go:64] FLAG: --v="2" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216047 4872 flags.go:64] FLAG: --version="false" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216053 4872 flags.go:64] FLAG: --vmodule="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216060 4872 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.216065 4872 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.216416 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220500 4872 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220508 4872 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220513 4872 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220519 4872 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220524 4872 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220528 4872 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220533 4872 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220537 4872 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220541 4872 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220545 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220550 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220554 4872 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220562 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220566 4872 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220571 4872 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220575 4872 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220605 4872 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220615 4872 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220619 4872 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220975 4872 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220986 4872 feature_gate.go:330] unrecognized feature gate: Example Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220991 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.220996 4872 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221000 4872 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221005 4872 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221010 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221015 4872 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221019 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221024 4872 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221028 4872 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221033 4872 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221037 4872 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221043 4872 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221048 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221053 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221057 4872 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221061 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221066 4872 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221070 4872 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221074 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221079 4872 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221083 4872 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221087 4872 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221091 4872 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221096 4872 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221100 4872 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221104 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221108 4872 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221114 4872 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221124 4872 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221129 4872 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221134 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221139 4872 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221144 4872 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221148 4872 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221153 4872 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221157 4872 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221161 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221165 4872 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221171 4872 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221175 4872 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221180 4872 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221183 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221187 4872 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221191 4872 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221197 4872 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221202 4872 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221206 4872 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221210 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.221216 4872 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.221223 4872 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.234602 4872 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.234692 4872 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234851 4872 feature_gate.go:330] unrecognized feature gate: Example Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234876 4872 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234886 4872 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234897 4872 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234908 4872 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234920 4872 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234933 4872 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234945 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234955 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234966 4872 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234976 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234985 4872 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.234995 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235003 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235012 4872 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235021 4872 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235029 4872 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235038 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235047 4872 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235056 4872 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235065 4872 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235073 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235081 4872 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235093 4872 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235103 4872 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235113 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235122 4872 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235131 4872 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235140 4872 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235149 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235159 4872 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235168 4872 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235177 4872 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235185 4872 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235193 4872 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235201 4872 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235210 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235218 4872 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235228 4872 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235239 4872 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235250 4872 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235259 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235268 4872 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235278 4872 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235289 4872 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235299 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235308 4872 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235318 4872 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235326 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235335 4872 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235343 4872 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235351 4872 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235360 4872 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235368 4872 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235376 4872 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235385 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235393 4872 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235402 4872 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235411 4872 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235420 4872 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235428 4872 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235438 4872 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235448 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235456 4872 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235465 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235473 4872 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235482 4872 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235490 4872 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235498 4872 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235506 4872 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235515 4872 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.235529 4872 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235803 4872 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235824 4872 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235836 4872 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235845 4872 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235854 4872 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235863 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235871 4872 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235880 4872 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235888 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235897 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235905 4872 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235913 4872 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235922 4872 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235931 4872 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235939 4872 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235948 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235956 4872 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235965 4872 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235975 4872 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235984 4872 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.235992 4872 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236001 4872 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236011 4872 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236020 4872 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236029 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236038 4872 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236046 4872 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236054 4872 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236063 4872 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236071 4872 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236079 4872 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236088 4872 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236098 4872 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236109 4872 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236119 4872 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236129 4872 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236138 4872 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236148 4872 feature_gate.go:330] unrecognized feature gate: Example Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236157 4872 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236165 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236174 4872 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236183 4872 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236191 4872 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236199 4872 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236208 4872 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236216 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236225 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236233 4872 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236242 4872 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236250 4872 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236259 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236268 4872 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236276 4872 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236284 4872 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236293 4872 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236302 4872 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236311 4872 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236319 4872 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236328 4872 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236336 4872 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236347 4872 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236357 4872 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236368 4872 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236378 4872 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236387 4872 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236396 4872 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236406 4872 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236415 4872 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236425 4872 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236433 4872 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.236441 4872 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.236455 4872 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.238716 4872 server.go:940] "Client rotation is on, will bootstrap in background" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.245300 4872 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.245468 4872 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.248746 4872 server.go:997] "Starting client certificate rotation" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.248802 4872 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.249038 4872 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-07 18:11:28.040485817 +0000 UTC Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.249119 4872 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 705h54m5.791370998s for next certificate rotation Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.275293 4872 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.278467 4872 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.293462 4872 log.go:25] "Validated CRI v1 runtime API" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.331344 4872 log.go:25] "Validated CRI v1 image API" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.334249 4872 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.340437 4872 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-09-08-12-46-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.340510 4872 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.367454 4872 manager.go:217] Machine: {Timestamp:2025-10-09 08:17:22.363923569 +0000 UTC m=+0.554452245 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:40bf4641-1076-4716-a2ae-ad6807813135 BootID:4a6bfc25-1aa1-4792-9216-2dff0eba4a5b Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:88:d1:86 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:88:d1:86 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:79:fc:15 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5d:97:18 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:97:ad:2d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:52:56:38 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:9a:e7:86:f8:d0:c8 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:16:8f:ab:13:8d:3e Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.367945 4872 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.368310 4872 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.369010 4872 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.369265 4872 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.369323 4872 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.371101 4872 topology_manager.go:138] "Creating topology manager with none policy" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.371137 4872 container_manager_linux.go:303] "Creating device plugin manager" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.371691 4872 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.371729 4872 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.371950 4872 state_mem.go:36] "Initialized new in-memory state store" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.372063 4872 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.375419 4872 kubelet.go:418] "Attempting to sync node with API server" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.375443 4872 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.375462 4872 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.375476 4872 kubelet.go:324] "Adding apiserver pod source" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.375490 4872 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.381018 4872 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.382283 4872 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.383634 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.383720 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.384604 4872 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.384759 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.384894 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386735 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386756 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386763 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386771 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386788 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386795 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386802 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386814 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386824 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386835 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386869 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.386877 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.388833 4872 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.389370 4872 server.go:1280] "Started kubelet" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.389585 4872 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.390281 4872 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.390963 4872 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.391006 4872 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 09 08:17:22 crc systemd[1]: Started Kubernetes Kubelet. Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.396970 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.397270 4872 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.402599 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 08:06:33.955816952 +0000 UTC Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.402666 4872 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1535h49m11.553154382s for next certificate rotation Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.403785 4872 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.403916 4872 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.403872 4872 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.403803 4872 server.go:460] "Adding debug handlers to kubelet server" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.405063 4872 factory.go:55] Registering systemd factory Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.405084 4872 factory.go:221] Registration of the systemd container factory successfully Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.402825 4872 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.405469 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="200ms" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.405748 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.405846 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.406011 4872 factory.go:153] Registering CRI-O factory Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.406051 4872 factory.go:221] Registration of the crio container factory successfully Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.406170 4872 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.406220 4872 factory.go:103] Registering Raw factory Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.406258 4872 manager.go:1196] Started watching for new ooms in manager Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.407782 4872 manager.go:319] Starting recovery of all containers Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.405065 4872 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186cc4b67911df65 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-09 08:17:22.389327717 +0000 UTC m=+0.579856363,LastTimestamp:2025-10-09 08:17:22.389327717 +0000 UTC m=+0.579856363,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416677 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416741 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416760 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416778 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416794 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416837 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416855 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416869 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416887 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416903 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416918 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416931 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416952 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.416974 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417002 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417019 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417038 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417053 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417071 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417107 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417123 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417139 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417157 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417238 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417258 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417281 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417299 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417318 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417334 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417351 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417367 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417388 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417406 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417423 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417440 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417457 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417475 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417495 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417512 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417530 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417548 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417566 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417585 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417605 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417624 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417665 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417687 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417706 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417725 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417774 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417793 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417813 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417840 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417862 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417881 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417900 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417919 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417938 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417953 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417972 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.417991 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418009 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418027 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418045 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418066 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418085 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418102 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418120 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418138 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418182 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418200 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418218 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418235 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418254 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418270 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418288 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418307 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418324 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418341 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418358 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418379 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418398 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418415 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418433 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418448 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418465 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418480 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418495 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418542 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418560 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418574 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418586 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418602 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418614 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418625 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418661 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418706 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418722 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418737 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418757 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418770 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.418782 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423317 4872 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423820 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423854 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423900 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423941 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423963 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.423990 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424017 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424049 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424069 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424095 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424116 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424140 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424159 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424185 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424201 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424222 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424237 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424259 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424284 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424300 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424324 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424398 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424417 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424439 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424458 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424479 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424495 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424515 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424539 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424568 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424621 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424671 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424697 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424722 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424775 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.424804 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425693 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425728 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425751 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425768 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425789 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425806 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425821 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425850 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425868 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425888 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425903 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425917 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425935 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425949 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425970 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425984 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.425998 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426019 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426037 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426057 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426071 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426085 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426103 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426119 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426136 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426150 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426164 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426187 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426202 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426222 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426238 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426251 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426272 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426286 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426301 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426320 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426334 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426353 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426374 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426390 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426408 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426421 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426438 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426450 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426463 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426482 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426498 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426519 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426533 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426547 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426567 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426582 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426604 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426618 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426631 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426666 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426680 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426696 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426712 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426726 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426747 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426764 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426785 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426802 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426818 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426837 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426852 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426873 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426889 4872 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426903 4872 reconstruct.go:97] "Volume reconstruction finished" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.426913 4872 reconciler.go:26] "Reconciler: start to sync state" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.427387 4872 manager.go:324] Recovery completed Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.444440 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.446287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.446356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.446375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.447771 4872 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.447808 4872 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.447870 4872 state_mem.go:36] "Initialized new in-memory state store" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.456005 4872 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.457904 4872 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.460283 4872 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.460358 4872 kubelet.go:2335] "Starting kubelet main sync loop" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.460434 4872 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.462260 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.462343 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.465432 4872 policy_none.go:49] "None policy: Start" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.466483 4872 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.466510 4872 state_mem.go:35] "Initializing new in-memory state store" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.504517 4872 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.527066 4872 manager.go:334] "Starting Device Plugin manager" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.527450 4872 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.527468 4872 server.go:79] "Starting device plugin registration server" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.528429 4872 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.528452 4872 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.528627 4872 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.528854 4872 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.528879 4872 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.539009 4872 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.561177 4872 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.561351 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.562774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.562826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.562838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.563045 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.563433 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.563503 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.563957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.563994 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564153 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564341 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564390 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564985 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.564996 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565147 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565291 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565329 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.565949 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.566075 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.566204 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.566205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.566307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.566323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.567997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568062 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568235 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.568265 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.569000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.569029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.569039 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.606490 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="400ms" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629290 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629353 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629523 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629554 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629576 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629593 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629612 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629628 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629737 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629810 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629857 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.629901 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.630296 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.630396 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.630487 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.630572 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.631656 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.631722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.631737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.631787 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.632335 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732000 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732057 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732076 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732097 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732114 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732131 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732183 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732201 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732220 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732236 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732300 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732317 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732332 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732401 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732580 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732575 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732691 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732727 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732786 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732805 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732827 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732813 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732831 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732847 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732877 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732895 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732896 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732948 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.732954 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.832924 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.835276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.835333 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.835347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.835379 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:22 crc kubenswrapper[4872]: E1009 08:17:22.835886 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.888590 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.907313 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.915489 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.930458 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: I1009 08:17:22.936163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.941921 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5213a8f73ce99ed8855bf6110fab84cb8ee432f5ea62938f3c55e250f2546b21 WatchSource:0}: Error finding container 5213a8f73ce99ed8855bf6110fab84cb8ee432f5ea62938f3c55e250f2546b21: Status 404 returned error can't find the container with id 5213a8f73ce99ed8855bf6110fab84cb8ee432f5ea62938f3c55e250f2546b21 Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.948723 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-ae0651df2affec8496028880b48d4993e5e01e95dc2ddb530205c46db2a17546 WatchSource:0}: Error finding container ae0651df2affec8496028880b48d4993e5e01e95dc2ddb530205c46db2a17546: Status 404 returned error can't find the container with id ae0651df2affec8496028880b48d4993e5e01e95dc2ddb530205c46db2a17546 Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.949829 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-cf80ea2bbf06ea2ce6274e3c3c63a2f12d6d0f214ab196ab73872527cb9d3042 WatchSource:0}: Error finding container cf80ea2bbf06ea2ce6274e3c3c63a2f12d6d0f214ab196ab73872527cb9d3042: Status 404 returned error can't find the container with id cf80ea2bbf06ea2ce6274e3c3c63a2f12d6d0f214ab196ab73872527cb9d3042 Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.956167 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-a6d16a5877439184d54a8ff6a17d5b8863369506d40425339714e1a98e452b2e WatchSource:0}: Error finding container a6d16a5877439184d54a8ff6a17d5b8863369506d40425339714e1a98e452b2e: Status 404 returned error can't find the container with id a6d16a5877439184d54a8ff6a17d5b8863369506d40425339714e1a98e452b2e Oct 09 08:17:22 crc kubenswrapper[4872]: W1009 08:17:22.961156 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-cdcddf9cc1e0476e49df150bcdc0d9f841c3a8c07549e5f28076115f57fde771 WatchSource:0}: Error finding container cdcddf9cc1e0476e49df150bcdc0d9f841c3a8c07549e5f28076115f57fde771: Status 404 returned error can't find the container with id cdcddf9cc1e0476e49df150bcdc0d9f841c3a8c07549e5f28076115f57fde771 Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.008168 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="800ms" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.236517 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.238071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.238122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.238137 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.238179 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.238972 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.392661 4872 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.466781 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"cdcddf9cc1e0476e49df150bcdc0d9f841c3a8c07549e5f28076115f57fde771"} Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.469445 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a6d16a5877439184d54a8ff6a17d5b8863369506d40425339714e1a98e452b2e"} Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.471423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cf80ea2bbf06ea2ce6274e3c3c63a2f12d6d0f214ab196ab73872527cb9d3042"} Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.473004 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ae0651df2affec8496028880b48d4993e5e01e95dc2ddb530205c46db2a17546"} Oct 09 08:17:23 crc kubenswrapper[4872]: I1009 08:17:23.474374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5213a8f73ce99ed8855bf6110fab84cb8ee432f5ea62938f3c55e250f2546b21"} Oct 09 08:17:23 crc kubenswrapper[4872]: W1009 08:17:23.581974 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.582081 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:23 crc kubenswrapper[4872]: W1009 08:17:23.754547 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.754709 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.809851 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="1.6s" Oct 09 08:17:23 crc kubenswrapper[4872]: W1009 08:17:23.834486 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.834684 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:23 crc kubenswrapper[4872]: W1009 08:17:23.962765 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:23 crc kubenswrapper[4872]: E1009 08:17:23.962869 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.039384 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.040978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.041036 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.041051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.041086 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:24 crc kubenswrapper[4872]: E1009 08:17:24.041836 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Oct 09 08:17:24 crc kubenswrapper[4872]: E1009 08:17:24.208670 4872 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.234:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186cc4b67911df65 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-09 08:17:22.389327717 +0000 UTC m=+0.579856363,LastTimestamp:2025-10-09 08:17:22.389327717 +0000 UTC m=+0.579856363,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.392142 4872 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.478444 4872 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="0c6e15d209b053323e9cdb636b99b1647206c080439e1693644571753f1b4166" exitCode=0 Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.478521 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"0c6e15d209b053323e9cdb636b99b1647206c080439e1693644571753f1b4166"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.478571 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.479593 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.479628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.479660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.481451 4872 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac" exitCode=0 Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.481518 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.481655 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.483508 4872 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7" exitCode=0 Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.483580 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.484163 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.484791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.484835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.484848 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.486446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.486654 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.486695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.490407 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.490455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.490469 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.490480 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.490583 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.491750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.491785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.491801 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493134 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca" exitCode=0 Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493173 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca"} Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493275 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.493992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.501860 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.518930 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.518981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:24 crc kubenswrapper[4872]: I1009 08:17:24.518992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.392482 4872 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:25 crc kubenswrapper[4872]: E1009 08:17:25.411334 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.234:6443: connect: connection refused" interval="3.2s" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.497953 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3f8ae3fb4ff28294fa996d1b34c20df9f9c146a1c8110f4fff78359fe37c68cf"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.498042 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.500038 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.500533 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.500548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.503026 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.503022 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.503173 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.503190 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.504792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.504859 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.504878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.512433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.512470 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.512483 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.512492 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.516165 4872 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3f8d642a3a4232e1f811e88c6bb2945ba6b5f35051c5c140e895200e2b6211f6" exitCode=0 Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.516249 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3f8d642a3a4232e1f811e88c6bb2945ba6b5f35051c5c140e895200e2b6211f6"} Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.516324 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.516392 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522370 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522649 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.522687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.572248 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.642280 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.643419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.643451 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.643460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.643482 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:25 crc kubenswrapper[4872]: E1009 08:17:25.643967 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.234:6443: connect: connection refused" node="crc" Oct 09 08:17:25 crc kubenswrapper[4872]: I1009 08:17:25.963474 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:25 crc kubenswrapper[4872]: W1009 08:17:25.965289 4872 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.234:6443: connect: connection refused Oct 09 08:17:25 crc kubenswrapper[4872]: E1009 08:17:25.965374 4872 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.234:6443: connect: connection refused" logger="UnhandledError" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.523402 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712"} Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.523566 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.525769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.525817 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.525831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527015 4872 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="61385e9bc0ebdd8029870ab831cb3f987f1eadee39cde00ae2ee71b2c740a4df" exitCode=0 Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527159 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527198 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527211 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"61385e9bc0ebdd8029870ab831cb3f987f1eadee39cde00ae2ee71b2c740a4df"} Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527278 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.527350 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529282 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529303 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529500 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:26 crc kubenswrapper[4872]: I1009 08:17:26.529437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.394179 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543219 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bd55878a609d70e447e63429c2dcf2310a5d141d0c0d73aa4839ee0e1382d3a0"} Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543336 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"be31a6aff4bdcd70a8ef42302a4c783557db3077a4778eb3b265ede9a3f7f2f6"} Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543364 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"acf935ba6bc652bcaafc59bdeb6af82b11c71981a53c7f46ec853b8cd595c424"} Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543489 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543613 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.543983 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.544633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.544743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.544762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.545470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.545499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.545513 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:27 crc kubenswrapper[4872]: I1009 08:17:27.653141 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.551273 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e6628c9a90887e474a0863cd6f568ac61c5363e854c87dd302ce683f64c61477"} Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.551346 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.551344 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d9ce3a5b9bb89177e19f1a3f5ddd916fc955e0fd5d09da6bbf07176d6fdfdf31"} Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.552188 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.552282 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.552787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.552818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.552829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.553805 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.844509 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.846351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.846392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.846407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:28 crc kubenswrapper[4872]: I1009 08:17:28.846437 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.553447 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.554507 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.554507 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.555966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.556044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.556084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.556757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.556807 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:29 crc kubenswrapper[4872]: I1009 08:17:29.556816 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.121321 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.395212 4872 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.395318 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.416017 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.557599 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.557599 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.558897 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.558959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.558977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.559057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.559090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:30 crc kubenswrapper[4872]: I1009 08:17:30.559100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.563556 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.565411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.565502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.565520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.596581 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.596862 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.598421 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.598458 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:31 crc kubenswrapper[4872]: I1009 08:17:31.598470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:32 crc kubenswrapper[4872]: E1009 08:17:32.539117 4872 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.825967 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.826230 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.827716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.827773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.827788 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:33 crc kubenswrapper[4872]: I1009 08:17:33.836882 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:34 crc kubenswrapper[4872]: I1009 08:17:34.570481 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:34 crc kubenswrapper[4872]: I1009 08:17:34.571693 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:34 crc kubenswrapper[4872]: I1009 08:17:34.571722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:34 crc kubenswrapper[4872]: I1009 08:17:34.571732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:34 crc kubenswrapper[4872]: I1009 08:17:34.574881 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:35 crc kubenswrapper[4872]: I1009 08:17:35.573266 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:35 crc kubenswrapper[4872]: I1009 08:17:35.574086 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:35 crc kubenswrapper[4872]: I1009 08:17:35.574143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:35 crc kubenswrapper[4872]: I1009 08:17:35.574164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:36 crc kubenswrapper[4872]: I1009 08:17:36.306067 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 09 08:17:36 crc kubenswrapper[4872]: I1009 08:17:36.306128 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 09 08:17:36 crc kubenswrapper[4872]: I1009 08:17:36.314098 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 09 08:17:36 crc kubenswrapper[4872]: I1009 08:17:36.314448 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.659805 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.660081 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.661871 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.661931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.661945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:37 crc kubenswrapper[4872]: I1009 08:17:37.664759 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:38 crc kubenswrapper[4872]: I1009 08:17:38.581803 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:38 crc kubenswrapper[4872]: I1009 08:17:38.583169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:38 crc kubenswrapper[4872]: I1009 08:17:38.583236 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:38 crc kubenswrapper[4872]: I1009 08:17:38.583264 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.592248 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.592448 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.593998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.594042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.594054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:39 crc kubenswrapper[4872]: I1009 08:17:39.611672 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.394988 4872 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.395062 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.586683 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.588175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.588241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:40 crc kubenswrapper[4872]: I1009 08:17:40.588267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.306236 4872 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.308345 4872 trace.go:236] Trace[1492693290]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 08:17:26.631) (total time: 14677ms): Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[1492693290]: ---"Objects listed" error: 14677ms (08:17:41.308) Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[1492693290]: [14.677283579s] [14.677283579s] END Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.308373 4872 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.311493 4872 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.311519 4872 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.312067 4872 trace.go:236] Trace[784705194]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 08:17:26.794) (total time: 14517ms): Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[784705194]: ---"Objects listed" error: 14517ms (08:17:41.311) Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[784705194]: [14.517848197s] [14.517848197s] END Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.312090 4872 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.312905 4872 trace.go:236] Trace[1190401508]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 08:17:30.511) (total time: 10799ms): Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[1190401508]: ---"Objects listed" error: 10799ms (08:17:41.311) Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[1190401508]: [10.799611994s] [10.799611994s] END Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.313243 4872 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.314107 4872 trace.go:236] Trace[269314966]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Oct-2025 08:17:26.961) (total time: 14352ms): Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[269314966]: ---"Objects listed" error: 14352ms (08:17:41.314) Oct 09 08:17:41 crc kubenswrapper[4872]: Trace[269314966]: [14.352215045s] [14.352215045s] END Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.314420 4872 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.332275 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41474->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.332307 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41490->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.333109 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41474->192.168.126.11:17697: read: connection reset by peer" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.333195 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:41490->192.168.126.11:17697: read: connection reset by peer" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.333526 4872 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.333559 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.387958 4872 apiserver.go:52] "Watching apiserver" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.391050 4872 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.391345 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.391855 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.391980 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.392041 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.392373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.392432 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.392439 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.392978 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.393028 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.393069 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.393410 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.393818 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.395039 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.395092 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.395050 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.395756 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.395862 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.398589 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.399049 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.406885 4872 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.411856 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.411902 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.411935 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.411962 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.411986 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412010 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412036 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412057 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412084 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412112 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412140 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412166 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412190 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412211 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412234 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412255 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412277 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412304 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412329 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412353 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412379 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412405 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412431 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412458 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412486 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412511 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412535 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412776 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412801 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412823 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412844 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412867 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412891 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412919 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412949 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412977 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413002 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413029 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413057 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413083 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413121 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413148 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413171 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413211 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413235 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413282 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413305 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413328 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413351 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413375 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413414 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413446 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413469 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413497 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413526 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413549 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413568 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413589 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413832 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413853 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413872 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413902 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413923 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413945 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413969 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.413993 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414015 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414038 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414059 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414081 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414103 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414125 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414148 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414173 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414197 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414220 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414244 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414269 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414294 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414318 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414342 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414365 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414391 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414417 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414441 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414462 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414485 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414505 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414528 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414553 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414574 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414599 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414622 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414688 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414712 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414735 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414758 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414782 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414806 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414829 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414851 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414876 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414900 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414925 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414945 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414966 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414989 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415013 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415070 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415099 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415122 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415146 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415169 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415193 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415215 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415235 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415283 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415304 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415325 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415348 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415372 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415396 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415420 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415442 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415466 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415491 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415614 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415676 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415704 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415729 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415752 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415776 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415799 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415824 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415854 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415875 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415897 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415919 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415942 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415966 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415991 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416015 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416038 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416062 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416085 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416112 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416137 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416161 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416186 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416211 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416237 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416287 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416315 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416357 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416384 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416419 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416443 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416467 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416493 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416515 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416546 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416573 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416597 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416626 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416672 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419963 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420082 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420135 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420180 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420226 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420272 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420312 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420355 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420388 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420425 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420460 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420498 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420527 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420566 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420607 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420634 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420685 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420722 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420753 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420787 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420822 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420856 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420885 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420923 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420957 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.420988 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421022 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421202 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421294 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421325 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421363 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421463 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421535 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421571 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421933 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421992 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.426464 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.432374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.434015 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.433258 4872 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.412943 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.414188 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415425 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415614 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.415795 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.434491 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416561 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416595 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419286 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419451 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419501 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419800 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.419895 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421374 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421630 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.421811 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.423216 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.423801 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.424012 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.424109 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.424272 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.424453 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.425157 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.426255 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.426821 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.427119 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.428696 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.428945 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.429276 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.429288 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.429539 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.416674 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.429808 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430092 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430126 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430181 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430205 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430378 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430450 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430481 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430866 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.430883 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431137 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431171 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431191 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431313 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431396 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431422 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431498 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431491 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431632 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431297 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431858 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.432049 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.432139 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.432248 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.431937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.432864 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.433070 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.433116 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.433239 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.433951 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.434185 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.434401 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.434779 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.435223 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.435256 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:41.935229234 +0000 UTC m=+20.125757860 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.435328 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:41.935319017 +0000 UTC m=+20.125847643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.435360 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.435583 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.435815 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.435818 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.436139 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.436185 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.436375 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.436486 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:17:41.936470262 +0000 UTC m=+20.126998888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.437039 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.437094 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.437444 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.437553 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.438060 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.438273 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.438430 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.438611 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.438715 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.439090 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.439148 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.439591 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.440017 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.440391 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.440683 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.442228 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.454632 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.455012 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.455173 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.455377 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.455872 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.456233 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.456623 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.457512 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.457937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.458247 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.458438 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.458616 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.459547 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.459586 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.459862 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.460543 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.460724 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.460966 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.461239 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.461297 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.462114 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.462511 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.462741 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.461630 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463031 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463113 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463290 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463822 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463905 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463939 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.463975 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464023 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464024 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464271 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464406 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464612 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.464820 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.465165 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.465198 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.465431 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.466586 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.465379 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.467493 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.468560 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.469383 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.469434 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.469585 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.469609 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.469825 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.470114 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.470184 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.474238 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.474361 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.474441 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.474585 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:41.974558279 +0000 UTC m=+20.165086905 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.475031 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.475309 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.475824 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.475849 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.475864 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:41 crc kubenswrapper[4872]: E1009 08:17:41.475924 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:41.97590126 +0000 UTC m=+20.166429886 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.480955 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.481178 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.481400 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.480764 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.484587 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.484824 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.484855 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.484947 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485071 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485375 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485534 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485462 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485464 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485734 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485748 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485939 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.485984 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486163 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486226 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486245 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486512 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486559 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.486931 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.487122 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.487801 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.488482 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.488521 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.488552 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.488914 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.489095 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490010 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-bd2z9"] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490238 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-7f65m"] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490220 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490587 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.490696 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.492883 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.492920 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.494832 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.498340 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.499500 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.499965 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500013 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500110 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500418 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500450 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500601 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500789 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.500876 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.501285 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.501870 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.502046 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.502351 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.503188 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.503738 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.503784 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.504565 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.504355 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.505085 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.505120 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.505365 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.507183 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.507412 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.507879 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.508063 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.508376 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.508675 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.509710 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.509944 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.514676 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.522756 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q8ss\" (UniqueName: \"kubernetes.io/projected/5e46110a-3d1a-431d-8489-1af8359725a4-kube-api-access-7q8ss\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523196 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523336 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4d45d7db-4c94-4b84-83b1-9049c7441fed-serviceca\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523391 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5e46110a-3d1a-431d-8489-1af8359725a4-hosts-file\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523466 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523561 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4d45d7db-4c94-4b84-83b1-9049c7441fed-host\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523586 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2c5\" (UniqueName: \"kubernetes.io/projected/4d45d7db-4c94-4b84-83b1-9049c7441fed-kube-api-access-ld2c5\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523685 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.523896 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524019 4872 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524055 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524068 4872 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524144 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524162 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524174 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524194 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524212 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524224 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524237 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524249 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524261 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524273 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524287 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524300 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524313 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524329 4872 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524419 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524485 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524502 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524513 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524526 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524411 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524538 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524671 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524688 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524701 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524714 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524724 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524754 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524764 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524773 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524782 4872 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524790 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524798 4872 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524811 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524839 4872 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524848 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524857 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524924 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524936 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524948 4872 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524962 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.524972 4872 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525010 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525022 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525034 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525045 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525056 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525091 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525105 4872 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525116 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525131 4872 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525141 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525193 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525211 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525223 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525238 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525282 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525295 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525307 4872 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525321 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525360 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525383 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525396 4872 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525408 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525698 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525716 4872 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525727 4872 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525758 4872 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525772 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525785 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525798 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525813 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525862 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525872 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525881 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525893 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525944 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525957 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525969 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.525982 4872 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526017 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526030 4872 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526043 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526055 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526068 4872 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526102 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526114 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526122 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526132 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526142 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526151 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526180 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526190 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526199 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526209 4872 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526219 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526229 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526239 4872 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526249 4872 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526263 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526272 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526281 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526289 4872 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526298 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526308 4872 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526316 4872 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526326 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526334 4872 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526343 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526353 4872 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526362 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526371 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526390 4872 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526398 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526408 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526416 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526425 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526434 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526443 4872 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526452 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526462 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526472 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526487 4872 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526500 4872 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526511 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526524 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526535 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526547 4872 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526556 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526569 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526578 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526589 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526600 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526609 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526618 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526627 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526636 4872 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526663 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526674 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526683 4872 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526694 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526704 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526715 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526739 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526748 4872 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526757 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526765 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526774 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526783 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526793 4872 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526802 4872 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526812 4872 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526821 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526830 4872 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526840 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526849 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526859 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526869 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526879 4872 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526889 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526899 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526910 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526919 4872 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526929 4872 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526938 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526947 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526956 4872 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526966 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526975 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526984 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.526995 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527004 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527013 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527023 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527032 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527040 4872 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527048 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527057 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527065 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527101 4872 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527114 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527185 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527196 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527205 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527226 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527236 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527264 4872 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527275 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.527299 4872 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.528907 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.532577 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.535467 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.538258 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.541181 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.554609 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.576551 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.587904 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.592158 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.594859 4872 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712" exitCode=255 Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.594904 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712"} Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.599294 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.610742 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.620443 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.627911 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4d45d7db-4c94-4b84-83b1-9049c7441fed-host\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.627988 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2c5\" (UniqueName: \"kubernetes.io/projected/4d45d7db-4c94-4b84-83b1-9049c7441fed-kube-api-access-ld2c5\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628022 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q8ss\" (UniqueName: \"kubernetes.io/projected/5e46110a-3d1a-431d-8489-1af8359725a4-kube-api-access-7q8ss\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628029 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/4d45d7db-4c94-4b84-83b1-9049c7441fed-host\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628040 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4d45d7db-4c94-4b84-83b1-9049c7441fed-serviceca\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628092 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5e46110a-3d1a-431d-8489-1af8359725a4-hosts-file\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628132 4872 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628143 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628153 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628163 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.628206 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/5e46110a-3d1a-431d-8489-1af8359725a4-hosts-file\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.629097 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/4d45d7db-4c94-4b84-83b1-9049c7441fed-serviceca\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.629116 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.636699 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.643682 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q8ss\" (UniqueName: \"kubernetes.io/projected/5e46110a-3d1a-431d-8489-1af8359725a4-kube-api-access-7q8ss\") pod \"node-resolver-bd2z9\" (UID: \"5e46110a-3d1a-431d-8489-1af8359725a4\") " pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.645742 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.647727 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2c5\" (UniqueName: \"kubernetes.io/projected/4d45d7db-4c94-4b84-83b1-9049c7441fed-kube-api-access-ld2c5\") pod \"node-ca-7f65m\" (UID: \"4d45d7db-4c94-4b84-83b1-9049c7441fed\") " pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.665486 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.683304 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.706690 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.715487 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.719905 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.735167 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.736910 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.754497 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.760497 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.760687 4872 scope.go:117] "RemoveContainer" containerID="cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.781573 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.798785 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.820491 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.822805 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-bd2z9" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.827984 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.828807 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-7f65m" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.856715 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-98kmz"] Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.857123 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.862583 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.862795 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.863007 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.863414 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 09 08:17:41 crc kubenswrapper[4872]: W1009 08:17:41.867401 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e46110a_3d1a_431d_8489_1af8359725a4.slice/crio-d316fd81e42b03afe90d59932b70a1534e0729a6d705c3482902bad5448772df WatchSource:0}: Error finding container d316fd81e42b03afe90d59932b70a1534e0729a6d705c3482902bad5448772df: Status 404 returned error can't find the container with id d316fd81e42b03afe90d59932b70a1534e0729a6d705c3482902bad5448772df Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.867925 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.889069 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.912196 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.921690 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.932004 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-mcd-auth-proxy-config\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.932068 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-rootfs\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.932094 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngxs4\" (UniqueName: \"kubernetes.io/projected/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-kube-api-access-ngxs4\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.932176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-proxy-tls\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.935859 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.944884 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.957109 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.972365 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.983423 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:41 crc kubenswrapper[4872]: I1009 08:17:41.997313 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.007149 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.032664 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.032776 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngxs4\" (UniqueName: \"kubernetes.io/projected/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-kube-api-access-ngxs4\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.032804 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.032859 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:17:43.032816533 +0000 UTC m=+21.223345159 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.032887 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.032940 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-proxy-tls\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.032950 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:43.032932066 +0000 UTC m=+21.223460682 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033032 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-mcd-auth-proxy-config\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033068 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033097 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033132 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-rootfs\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033170 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033247 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033275 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033287 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033340 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:43.033319698 +0000 UTC m=+21.223848324 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033342 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033381 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:43.03337562 +0000 UTC m=+21.223904236 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.033427 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-rootfs\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033434 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033460 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033476 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.033530 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:43.033515534 +0000 UTC m=+21.224044340 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.034052 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-mcd-auth-proxy-config\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.036568 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-proxy-tls\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.048886 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngxs4\" (UniqueName: \"kubernetes.io/projected/5cc7b6a1-93b1-455b-aa21-d8bb7574857e-kube-api-access-ngxs4\") pod \"machine-config-daemon-98kmz\" (UID: \"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\") " pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.174377 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.221550 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-ssrfz"] Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.222295 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-5ftzb"] Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.222556 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.222916 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.224936 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xnhdk"] Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.225618 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.226505 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.226841 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.227167 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233234 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233488 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233549 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233766 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233845 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.233991 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.234327 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.234555 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.236055 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.237075 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.237275 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.243481 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.255477 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.276956 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.291065 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.306598 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.318247 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.331933 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336426 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336511 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336566 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336587 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336668 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-system-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336693 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336739 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336758 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-daemon-config\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336824 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-multus\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336846 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-hostroot\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336905 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-os-release\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336932 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.336993 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cni-binary-copy\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337058 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-socket-dir-parent\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-k8s-cni-cncf-io\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337174 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337202 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337229 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-conf-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337256 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337303 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337333 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337401 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337458 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssc5b\" (UniqueName: \"kubernetes.io/projected/d5f349b9-7850-4b1b-86da-e89f7cfd32df-kube-api-access-ssc5b\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337478 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snww9\" (UniqueName: \"kubernetes.io/projected/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-kube-api-access-snww9\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337550 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337578 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-multus-certs\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337597 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337630 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337682 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337707 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-kubelet\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337838 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-system-cni-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337863 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.337907 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-os-release\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338119 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cnibin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338166 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-netns\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338184 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cnibin\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338216 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338251 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85kdn\" (UniqueName: \"kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338298 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338327 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-bin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.338349 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-etc-kubernetes\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.345746 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.357080 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.365763 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.377519 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.391366 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.420570 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.432565 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439306 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439325 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439344 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439364 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-system-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439393 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439414 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439445 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-multus\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439465 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-hostroot\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439484 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-daemon-config\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439502 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-os-release\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439531 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-k8s-cni-cncf-io\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439567 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-system-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439593 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439699 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cni-binary-copy\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439779 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-socket-dir-parent\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439829 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439852 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-conf-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439872 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439891 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439927 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snww9\" (UniqueName: \"kubernetes.io/projected/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-kube-api-access-snww9\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439947 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439980 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssc5b\" (UniqueName: \"kubernetes.io/projected/d5f349b9-7850-4b1b-86da-e89f7cfd32df-kube-api-access-ssc5b\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.439999 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440015 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440033 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440051 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-multus-certs\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440069 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-system-cni-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440084 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440101 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440129 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-kubelet\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440151 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-os-release\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440167 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440191 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-netns\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440207 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cnibin\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440223 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440240 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440260 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cnibin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440253 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440285 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-bin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440361 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-etc-kubernetes\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440375 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-binary-copy\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85kdn\" (UniqueName: \"kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440553 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440585 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-cni-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440586 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440609 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-etc-kubernetes\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440621 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-socket-dir-parent\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440673 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440674 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440673 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-k8s-cni-cncf-io\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440698 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440732 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440742 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440767 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440751 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-hostroot\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440735 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-multus\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440769 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cni-binary-copy\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440794 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440812 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-cnibin\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440783 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-kubelet\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440797 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440902 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-conf-dir\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440902 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440922 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440965 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-system-cni-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440995 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.440993 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441012 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/d5f349b9-7850-4b1b-86da-e89f7cfd32df-multus-daemon-config\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441019 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-multus-certs\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441047 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441055 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441074 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-var-lib-cni-bin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441111 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-cnibin\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441144 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-host-run-netns\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441169 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d5f349b9-7850-4b1b-86da-e89f7cfd32df-os-release\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441276 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-os-release\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441360 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.441445 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-tuning-conf-dir\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.452606 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.462137 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:42 crc kubenswrapper[4872]: E1009 08:17:42.462315 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.470359 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.471709 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.472470 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.473629 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.474304 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.476079 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.476654 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.477343 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.478836 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.479517 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.484609 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssc5b\" (UniqueName: \"kubernetes.io/projected/d5f349b9-7850-4b1b-86da-e89f7cfd32df-kube-api-access-ssc5b\") pod \"multus-5ftzb\" (UID: \"d5f349b9-7850-4b1b-86da-e89f7cfd32df\") " pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.485063 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.488234 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85kdn\" (UniqueName: \"kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn\") pod \"ovnkube-node-xnhdk\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.488812 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.489414 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.493507 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.496358 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snww9\" (UniqueName: \"kubernetes.io/projected/7bc7b555-0389-4f9c-9bc9-02387eb2f7c3-kube-api-access-snww9\") pod \"multus-additional-cni-plugins-ssrfz\" (UID: \"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\") " pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.498148 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.500080 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.500707 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.501802 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.502521 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.503028 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.504093 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.504793 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.505274 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.508308 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.509098 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.509764 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.511027 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.511601 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.512832 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.513523 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.514035 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.516728 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.517370 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.518419 4872 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.518544 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.520292 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.521242 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.521865 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.524980 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.527622 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.528697 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.529913 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.532346 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.532829 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.533671 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.535979 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.542333 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.543340 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.545132 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.546137 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.546740 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.548610 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.549275 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.550549 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.551050 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.551906 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.554230 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.554502 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.555244 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.556319 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5ftzb" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.562034 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.569211 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:42 crc kubenswrapper[4872]: W1009 08:17:42.590624 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7bc7b555_0389_4f9c_9bc9_02387eb2f7c3.slice/crio-e40dede807c545e4fe4ee6c6e12c934f373389745612998a53848c5bd1d75f6e WatchSource:0}: Error finding container e40dede807c545e4fe4ee6c6e12c934f373389745612998a53848c5bd1d75f6e: Status 404 returned error can't find the container with id e40dede807c545e4fe4ee6c6e12c934f373389745612998a53848c5bd1d75f6e Oct 09 08:17:42 crc kubenswrapper[4872]: W1009 08:17:42.593937 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfac60b03_54bb_43e8_8994_3674d42b209b.slice/crio-1e57250e14b035a83086e2213c2227f1cf0afb4cbf994c4184bbaad1bdf1f7be WatchSource:0}: Error finding container 1e57250e14b035a83086e2213c2227f1cf0afb4cbf994c4184bbaad1bdf1f7be: Status 404 returned error can't find the container with id 1e57250e14b035a83086e2213c2227f1cf0afb4cbf994c4184bbaad1bdf1f7be Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.597120 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.601044 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerStarted","Data":"e40dede807c545e4fe4ee6c6e12c934f373389745612998a53848c5bd1d75f6e"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.603335 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c2ec75bbc9ee3c4a29429ef08bebadf25fed2d283c2de8d30a3f6bd5f75beb24"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.610156 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.610186 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d5867a19b69009c041b20470b66fc25c2738eb60dd5520f20624b1decc3fe68d"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.612043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"1e57250e14b035a83086e2213c2227f1cf0afb4cbf994c4184bbaad1bdf1f7be"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.614929 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.614980 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"00b11627fc21fbda3f5c733fbbd90099552533d42f6b6215bb07de9e6b727144"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.617056 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bd2z9" event={"ID":"5e46110a-3d1a-431d-8489-1af8359725a4","Type":"ContainerStarted","Data":"acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.617089 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-bd2z9" event={"ID":"5e46110a-3d1a-431d-8489-1af8359725a4","Type":"ContainerStarted","Data":"d316fd81e42b03afe90d59932b70a1534e0729a6d705c3482902bad5448772df"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.619170 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.619197 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.619207 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"877946f3915daba6e142e956e5304c7bc9bc51eecedb6d541f0fe998b362728e"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.623837 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7f65m" event={"ID":"4d45d7db-4c94-4b84-83b1-9049c7441fed","Type":"ContainerStarted","Data":"a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.623906 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-7f65m" event={"ID":"4d45d7db-4c94-4b84-83b1-9049c7441fed","Type":"ContainerStarted","Data":"0458b8813217413e7e2bb41435e2e51b4a4965e7d20ab4086d6f50e847aeef04"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.626983 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerStarted","Data":"2db214f9036f3ab77e2045697f63b149dc3117cab756d0763577afd5ad460904"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.631683 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.634125 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1"} Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.634458 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.636315 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.682423 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.718122 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.763848 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.802466 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.838445 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.885254 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.916784 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:42 crc kubenswrapper[4872]: I1009 08:17:42.952860 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.000449 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.036261 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.045686 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.045790 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.045814 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.045832 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.045857 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046020 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:17:45.046005666 +0000 UTC m=+23.236534292 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046061 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046090 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:45.046084228 +0000 UTC m=+23.236612854 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046367 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046395 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046406 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046428 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:45.046422229 +0000 UTC m=+23.236950855 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046468 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046489 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:45.046483841 +0000 UTC m=+23.237012467 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046501 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046510 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046517 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.046537 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:45.046531422 +0000 UTC m=+23.237060048 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.079907 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.117136 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.159459 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.196433 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.239364 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.277633 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.317115 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.363186 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.395836 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.438119 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.461123 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.461143 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.461237 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:43 crc kubenswrapper[4872]: E1009 08:17:43.461397 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.482200 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.521417 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.556413 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.597553 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.639978 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.639996 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerStarted","Data":"956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90"} Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.641028 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" exitCode=0 Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.641095 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.643809 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e"} Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.646005 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78" exitCode=0 Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.646529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78"} Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.682335 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.717214 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.758400 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.796567 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.842418 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.899183 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.925143 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:43 crc kubenswrapper[4872]: I1009 08:17:43.965129 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:43Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.005530 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.039381 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.075853 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.123658 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.162419 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.204288 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.236958 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.286024 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.317896 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.355536 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.461299 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:44 crc kubenswrapper[4872]: E1009 08:17:44.461466 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655184 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655521 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655585 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655687 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655783 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.655865 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.657629 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588" exitCode=0 Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.657678 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588"} Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.685483 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.700016 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.715017 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.729764 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.743747 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.757445 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.773164 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.786133 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.804315 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.836873 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.857014 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.871461 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:44 crc kubenswrapper[4872]: I1009 08:17:44.886370 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:44Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.067306 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067526 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:17:49.067504895 +0000 UTC m=+27.258033521 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.067594 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.067630 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.067702 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.067763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067858 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067865 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067899 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067917 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067919 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067903 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:49.067896347 +0000 UTC m=+27.258424973 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067968 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.067992 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.068002 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:49.06798097 +0000 UTC m=+27.258509786 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.068060 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:49.068038261 +0000 UTC m=+27.258566987 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.068056 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.068219 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:49.068191506 +0000 UTC m=+27.258720152 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.461224 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.461309 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.461370 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:45 crc kubenswrapper[4872]: E1009 08:17:45.461457 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.663290 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983" exitCode=0 Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.663355 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983"} Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.665423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e"} Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.678559 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.690578 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.703862 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.717179 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.734826 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.745968 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.760451 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.772045 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.787475 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.805060 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.824336 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.845103 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.858867 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.874088 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.887144 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.900901 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.916286 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.930757 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.944311 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.968495 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.981000 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:45 crc kubenswrapper[4872]: I1009 08:17:45.990590 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:45Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.004319 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.017540 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.030975 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.044154 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.461819 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:46 crc kubenswrapper[4872]: E1009 08:17:46.461995 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.671241 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060" exitCode=0 Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.671279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060"} Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.684502 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.705287 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.716472 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.732330 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.749869 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.767100 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.780664 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.796041 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.810289 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.823944 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.838511 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.853007 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:46 crc kubenswrapper[4872]: I1009 08:17:46.864240 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.400871 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.405979 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.413894 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.417914 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.431430 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.448297 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.461188 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.461287 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.461710 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.461720 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.465766 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.488771 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.503347 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.523086 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.540331 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.558941 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.571375 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.587234 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.603831 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.619337 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.632385 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.648176 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.670372 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.679011 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.682017 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76" exitCode=0 Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.682253 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.696040 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.708260 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.715291 4872 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.717539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.717576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.717590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.717777 4872 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.723799 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.725433 4872 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.725795 4872 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.726808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.726846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.726863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.726886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.726903 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.740038 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.740599 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.744496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.744527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.744539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.744559 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.744574 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.754274 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.758328 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.763091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.763129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.763142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.763161 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.763173 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.767526 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.778230 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.778946 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.782604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.782667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.782678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.782694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.782704 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.795755 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.798465 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.807471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.807516 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.807530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.807581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.807597 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.809695 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.822178 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: E1009 08:17:47.822305 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.822460 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.824306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.824430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.824444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.824460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.824471 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.835905 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.848477 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.866102 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.891805 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.904483 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.915421 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.926577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.926621 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.926632 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.926664 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.926677 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:47Z","lastTransitionTime":"2025-10-09T08:17:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.927129 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.937006 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.947848 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.958382 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:47 crc kubenswrapper[4872]: I1009 08:17:47.976197 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.022349 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.028489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.028561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.028587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.028618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.028672 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.059096 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.099198 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.131037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.131097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.131116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.131145 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.131166 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.138427 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.234165 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.234217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.234229 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.234249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.234260 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.337726 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.337766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.337775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.337792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.337802 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.440399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.440480 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.440494 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.440514 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.440530 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.461129 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:48 crc kubenswrapper[4872]: E1009 08:17:48.461388 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.543040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.543097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.543114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.543139 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.543156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.652372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.652442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.652460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.652493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.652532 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.691735 4872 generic.go:334] "Generic (PLEG): container finished" podID="7bc7b555-0389-4f9c-9bc9-02387eb2f7c3" containerID="115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf" exitCode=0 Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.691796 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerDied","Data":"115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.713462 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.734186 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.750718 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.756999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.757065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.757082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.757111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.757132 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.770470 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.787667 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.800371 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.813682 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.830470 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.844960 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.858866 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.860890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.860928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.860940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.860958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.860969 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.868350 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.885724 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.909464 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.922494 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.964218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.964264 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.964273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.964289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:48 crc kubenswrapper[4872]: I1009 08:17:48.964299 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:48Z","lastTransitionTime":"2025-10-09T08:17:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.067671 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.067732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.067745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.067766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.067779 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.111016 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.110989406 +0000 UTC m=+35.301518082 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.110858 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.111513 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.111551 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.111579 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.111711 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.111684788 +0000 UTC m=+35.302213464 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.111283 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.111979 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.112062 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.112151 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112210 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112288 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112364 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.112340208 +0000 UTC m=+35.302868874 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112441 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112476 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112499 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112590 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.112565985 +0000 UTC m=+35.303094681 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.112960 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.112934226 +0000 UTC m=+35.303462862 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.170085 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.170128 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.170141 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.170157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.170168 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.273281 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.273340 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.273355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.273377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.273393 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.376022 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.376056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.376067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.376082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.376093 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.461618 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.461660 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.461756 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:49 crc kubenswrapper[4872]: E1009 08:17:49.461866 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.478796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.478823 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.478832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.478847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.478856 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.581208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.581244 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.581253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.581267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.581278 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.684231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.684307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.684326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.684352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.684376 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.701792 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.705835 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.705919 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.716918 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" event={"ID":"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3","Type":"ContainerStarted","Data":"2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.726683 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.743531 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.762231 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.776356 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.787539 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.787678 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.787716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.787763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.787781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.788201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.788264 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.796905 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.812060 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.832142 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.846573 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.860779 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.871614 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.882162 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.891475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.891523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.891537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.891558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.891572 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.892207 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.906939 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.926720 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.940625 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.952093 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.963909 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.975008 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.993632 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.993698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.993710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.993728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:49 crc kubenswrapper[4872]: I1009 08:17:49.993740 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:49Z","lastTransitionTime":"2025-10-09T08:17:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.009354 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.043838 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.060111 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.071886 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.081068 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.092574 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.095439 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.095475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.095484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.095500 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.095509 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.103394 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.113905 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.123690 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.135617 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:50Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.197937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.197989 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.198018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.198033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.198042 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.300525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.300589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.300606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.300635 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.300689 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.403449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.403509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.403527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.403551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.403569 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.461461 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:50 crc kubenswrapper[4872]: E1009 08:17:50.461670 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.506716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.506787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.506813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.506847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.506871 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.609957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.610023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.610039 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.610061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.610075 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.712698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.712747 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.712762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.712783 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.712793 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.720503 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.816212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.816361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.816385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.816415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.816440 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.919727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.919790 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.919811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.919838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:50 crc kubenswrapper[4872]: I1009 08:17:50.919858 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:50Z","lastTransitionTime":"2025-10-09T08:17:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.023050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.023108 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.023126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.023152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.023169 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.125420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.125462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.125476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.125532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.125546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.227639 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.227703 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.227718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.227732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.227740 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.330538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.330598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.330615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.330634 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.330670 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.433713 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.433762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.433771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.433788 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.433797 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.461160 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.461174 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:51 crc kubenswrapper[4872]: E1009 08:17:51.461289 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:51 crc kubenswrapper[4872]: E1009 08:17:51.461410 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.536025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.536131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.536153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.536181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.536202 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.639410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.639457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.639472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.639493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.639508 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.723762 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.741806 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.741842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.741852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.741867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.741877 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.846032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.846067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.846078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.846095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.846106 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.948786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.948836 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.948849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.948868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:51 crc kubenswrapper[4872]: I1009 08:17:51.948886 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:51Z","lastTransitionTime":"2025-10-09T08:17:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.052604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.052692 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.052707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.052730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.052772 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.156224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.156323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.156341 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.156396 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.156415 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.260091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.260175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.260212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.260242 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.260260 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.369596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.369734 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.369760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.369832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.369855 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.461703 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:52 crc kubenswrapper[4872]: E1009 08:17:52.461901 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.473168 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.473216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.473231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.473250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.473266 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.487441 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.508752 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.525751 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.546370 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.567340 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.575442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.575494 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.575507 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.575527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.575539 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.584256 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.596719 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.610991 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.626698 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.639283 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.651546 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.663221 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.677622 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.678559 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.678596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.678605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.678620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.678630 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.696246 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.728950 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/0.log" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.732495 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17" exitCode=1 Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.732539 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.733388 4872 scope.go:117] "RemoveContainer" containerID="8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.758733 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.772849 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.781903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.781955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.781972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.781998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.782015 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.784815 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.805155 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.823619 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.841581 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.859137 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.873724 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.885129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.885182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.885198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.885219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.885234 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.894990 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.909238 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.925949 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.937542 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.961954 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.989783 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.989847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.989867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.989897 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.989917 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:52Z","lastTransitionTime":"2025-10-09T08:17:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:52 crc kubenswrapper[4872]: I1009 08:17:52.992443 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.092575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.092617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.092627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.092655 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.092668 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.174525 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.187051 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.195620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.195694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.195711 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.195737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.195755 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.200629 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.211152 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.225397 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.250959 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.267764 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.282322 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.298420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.298467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.298479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.298497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.298508 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.302048 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.313242 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.325894 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.339839 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.349458 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.366022 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.382296 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.401219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.401324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.401336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.401358 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.401374 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.460928 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.461043 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:53 crc kubenswrapper[4872]: E1009 08:17:53.461087 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:53 crc kubenswrapper[4872]: E1009 08:17:53.461292 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.503903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.503986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.503998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.504018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.504029 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.606262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.606316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.606328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.606344 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.606355 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.708407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.708458 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.708470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.708488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.708502 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.739122 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/0.log" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.743456 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.743762 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.758681 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.770612 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.790393 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.811814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.811869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.811889 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.811915 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.811934 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.815935 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.848980 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.863450 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.874938 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.890124 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.910499 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.914737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.914775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.914786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.914804 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.914816 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:53Z","lastTransitionTime":"2025-10-09T08:17:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.931959 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.947892 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.967026 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:53 crc kubenswrapper[4872]: I1009 08:17:53.986941 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:53Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.003629 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.017695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.017732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.017740 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.017754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.017762 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.120383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.120457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.120473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.120496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.120512 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.223774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.223823 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.223834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.223853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.223864 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.326717 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.326793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.326818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.326848 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.326871 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.429576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.429693 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.429720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.429748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.429773 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.461526 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:54 crc kubenswrapper[4872]: E1009 08:17:54.461776 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.533486 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.533553 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.533577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.533608 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.533633 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.578208 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs"] Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.579589 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.582692 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.582854 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.601478 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.620591 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.637330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.637467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.637495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.637523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.637541 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.643385 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.658022 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.676378 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.682799 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.682846 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.682866 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvbxk\" (UniqueName: \"kubernetes.io/projected/63b22313-f1d5-44b5-877b-e5aca8937437-kube-api-access-bvbxk\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.682936 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63b22313-f1d5-44b5-877b-e5aca8937437-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.694118 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.707807 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.722857 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.741138 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.741334 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.741388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.741418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.741470 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.744886 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.749082 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/1.log" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.750450 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/0.log" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.754520 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1" exitCode=1 Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.754575 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.754675 4872 scope.go:117] "RemoveContainer" containerID="8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.756044 4872 scope.go:117] "RemoveContainer" containerID="2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1" Oct 09 08:17:54 crc kubenswrapper[4872]: E1009 08:17:54.756458 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.767417 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.783707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63b22313-f1d5-44b5-877b-e5aca8937437-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.783910 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.784027 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvbxk\" (UniqueName: \"kubernetes.io/projected/63b22313-f1d5-44b5-877b-e5aca8937437-kube-api-access-bvbxk\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.784139 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.785170 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-env-overrides\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.785628 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/63b22313-f1d5-44b5-877b-e5aca8937437-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.789732 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.791909 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/63b22313-f1d5-44b5-877b-e5aca8937437-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.803488 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvbxk\" (UniqueName: \"kubernetes.io/projected/63b22313-f1d5-44b5-877b-e5aca8937437-kube-api-access-bvbxk\") pod \"ovnkube-control-plane-749d76644c-s7ffs\" (UID: \"63b22313-f1d5-44b5-877b-e5aca8937437\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.813087 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.830008 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.845386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.845455 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.845473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.845501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.845520 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.846198 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.863285 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.879299 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.891368 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.895165 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.906854 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: W1009 08:17:54.913989 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod63b22313_f1d5_44b5_877b_e5aca8937437.slice/crio-876962177e066017830d6aa3f8594cbc8c86b6dfea199926fcc701f9cdadb775 WatchSource:0}: Error finding container 876962177e066017830d6aa3f8594cbc8c86b6dfea199926fcc701f9cdadb775: Status 404 returned error can't find the container with id 876962177e066017830d6aa3f8594cbc8c86b6dfea199926fcc701f9cdadb775 Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.938714 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.948407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.948463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.948483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.948512 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.948530 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:54Z","lastTransitionTime":"2025-10-09T08:17:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.961088 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.977898 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:54 crc kubenswrapper[4872]: I1009 08:17:54.990272 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:54Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.010169 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.026539 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.043434 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.052092 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.052153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.052165 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.052207 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.052223 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.055680 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.072923 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.087463 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.099232 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.110883 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.154534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.154581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.154592 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.154610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.154621 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.257976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.258036 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.258054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.258080 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.258098 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.360976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.361037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.361051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.361079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.361095 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.461332 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.461440 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:55 crc kubenswrapper[4872]: E1009 08:17:55.461445 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:55 crc kubenswrapper[4872]: E1009 08:17:55.461578 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.463479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.463519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.463529 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.463545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.463556 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.565984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.566035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.566051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.566072 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.566084 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.669924 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.670007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.670020 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.670043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.670058 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.675114 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-8grb6"] Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.675967 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: E1009 08:17:55.676079 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.695835 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.716373 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.736683 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.762166 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" event={"ID":"63b22313-f1d5-44b5-877b-e5aca8937437","Type":"ContainerStarted","Data":"0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.762274 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" event={"ID":"63b22313-f1d5-44b5-877b-e5aca8937437","Type":"ContainerStarted","Data":"32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.762308 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" event={"ID":"63b22313-f1d5-44b5-877b-e5aca8937437","Type":"ContainerStarted","Data":"876962177e066017830d6aa3f8594cbc8c86b6dfea199926fcc701f9cdadb775"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.764510 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/1.log" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.767490 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.772704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.772825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.772964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.773056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.773145 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.781328 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.796188 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jczsj\" (UniqueName: \"kubernetes.io/projected/326d4497-ac5f-4a64-afde-14ded51665b8-kube-api-access-jczsj\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.796251 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.803788 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.829810 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.846763 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.863582 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.877114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.877209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.877223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.877323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.877336 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.878131 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.891348 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.897531 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jczsj\" (UniqueName: \"kubernetes.io/projected/326d4497-ac5f-4a64-afde-14ded51665b8-kube-api-access-jczsj\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.897628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: E1009 08:17:55.897886 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:55 crc kubenswrapper[4872]: E1009 08:17:55.897956 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:56.397936662 +0000 UTC m=+34.588465298 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.905682 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.919110 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jczsj\" (UniqueName: \"kubernetes.io/projected/326d4497-ac5f-4a64-afde-14ded51665b8-kube-api-access-jczsj\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.922090 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.936235 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.948101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.962136 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.973394 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.979238 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.979282 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.979294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.979312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.979323 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:55Z","lastTransitionTime":"2025-10-09T08:17:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.985430 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:55 crc kubenswrapper[4872]: I1009 08:17:55.997756 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:55Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.006151 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.018848 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.037347 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8df79cad04d5bd1b501064d81900f77fc5d2c3d9c837acef0ecd4c6f5508ce17\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:51Z\\\",\\\"message\\\":\\\"1.811024 6174 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1009 08:17:51.811030 6174 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:51.811040 6174 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:51.811137 6174 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811522 6174 factory.go:656] Stopping watch factory\\\\nI1009 08:17:51.811523 6174 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811574 6174 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811668 6174 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1009 08:17:51.811713 6174 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:51.811849 6174 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.047891 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.056952 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.065395 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.080711 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.082174 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.082240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.082258 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.082286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.082305 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.095397 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.109718 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.123210 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.136386 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.156861 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185440 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:56Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185612 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.185623 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.288305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.288359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.288368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.288386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.288408 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.391753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.391811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.391824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.391846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.391859 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.401554 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:56 crc kubenswrapper[4872]: E1009 08:17:56.401746 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:56 crc kubenswrapper[4872]: E1009 08:17:56.401862 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:57.401840802 +0000 UTC m=+35.592369528 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.461190 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:56 crc kubenswrapper[4872]: E1009 08:17:56.462504 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.494028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.494098 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.494117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.494143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.494161 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.597837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.597920 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.597939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.597968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.597989 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.701294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.701392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.701416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.701444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.701464 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.805476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.805547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.805560 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.805581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.805594 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.908032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.908589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.908613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.908684 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:56 crc kubenswrapper[4872]: I1009 08:17:56.908711 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:56Z","lastTransitionTime":"2025-10-09T08:17:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.011259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.011306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.011317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.011336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.011348 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.114075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.114142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.114155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.114174 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.114189 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.209038 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.209195 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.209256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.209293 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.209343 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209448 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209518 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209551 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209587 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209611 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209759 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209819 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209848 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209511 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:18:13.209450179 +0000 UTC m=+51.399978845 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.209965 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:13.209939624 +0000 UTC m=+51.400468290 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.210023 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:13.210004586 +0000 UTC m=+51.400533332 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.210053 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:13.210038348 +0000 UTC m=+51.400567004 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.210092 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:13.210078059 +0000 UTC m=+51.400606775 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.218299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.218377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.218397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.218421 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.218443 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.320800 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.320839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.320850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.320867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.320878 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.410763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.410966 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.411050 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:17:59.411027352 +0000 UTC m=+37.601556018 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.423996 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.424107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.424129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.424156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.424176 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.460981 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.461238 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.460981 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.461472 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.461776 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:57 crc kubenswrapper[4872]: E1009 08:17:57.462085 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.527070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.527111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.527127 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.527152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.527194 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.630384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.630456 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.630479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.630510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.630531 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.733409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.733448 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.733457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.733474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.733483 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.836084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.836142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.836159 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.836182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.836204 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.939481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.939547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.939565 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.939591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:57 crc kubenswrapper[4872]: I1009 08:17:57.939610 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:57Z","lastTransitionTime":"2025-10-09T08:17:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.043066 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.043130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.043150 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.043175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.043193 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.098198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.098272 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.098293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.098317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.098335 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.120058 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:58Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.125084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.125163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.125187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.125221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.125249 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.146635 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:58Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.151461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.151532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.151550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.151578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.151598 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.177834 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:58Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.183028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.183106 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.183131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.183164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.183189 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.201211 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:58Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.206361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.206424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.206444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.206471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.206489 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.229231 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:17:58Z is after 2025-08-24T17:21:41Z" Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.229453 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.232103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.232196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.232215 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.232243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.232261 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.335689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.335778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.335798 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.335833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.335852 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.438572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.438699 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.438718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.438743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.438759 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.461134 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:17:58 crc kubenswrapper[4872]: E1009 08:17:58.461373 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.542018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.542087 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.542111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.542144 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.542165 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.645905 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.645960 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.645977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.646002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.646020 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.748954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.749023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.749039 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.749063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.749079 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.852426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.852498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.852522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.852556 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.852581 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.956385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.956466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.956490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.956523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:58 crc kubenswrapper[4872]: I1009 08:17:58.956544 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:58Z","lastTransitionTime":"2025-10-09T08:17:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.059809 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.059856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.059865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.059881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.059890 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.163400 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.163469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.163487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.163517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.163535 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.266449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.266509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.266521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.266540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.266557 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.370508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.370560 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.370577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.370602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.370618 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.434025 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:59 crc kubenswrapper[4872]: E1009 08:17:59.434322 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:59 crc kubenswrapper[4872]: E1009 08:17:59.434425 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:03.434395798 +0000 UTC m=+41.624924454 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.461053 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.461090 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.461053 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:17:59 crc kubenswrapper[4872]: E1009 08:17:59.461237 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:17:59 crc kubenswrapper[4872]: E1009 08:17:59.461385 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:17:59 crc kubenswrapper[4872]: E1009 08:17:59.461556 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.474271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.474335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.474391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.474417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.474435 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.578036 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.578112 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.578130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.578155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.578177 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.681376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.681422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.681433 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.681450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.681461 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.783714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.783780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.783803 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.783833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.783855 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.886503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.886585 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.886611 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.886688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.886715 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.990053 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.990143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.990162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.990185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:17:59 crc kubenswrapper[4872]: I1009 08:17:59.990201 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:17:59Z","lastTransitionTime":"2025-10-09T08:17:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.092999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.093040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.093049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.093063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.093072 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.195824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.195894 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.195916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.195948 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.195972 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.298558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.298633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.298712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.298743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.298763 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.402025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.402090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.402114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.402147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.402171 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.460817 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:00 crc kubenswrapper[4872]: E1009 08:18:00.461049 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.505160 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.505200 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.505214 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.505230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.505241 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.608610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.608719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.608740 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.608763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.608780 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.711366 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.711430 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.711447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.711477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.711495 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.814905 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.814991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.815021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.815051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.815072 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.919203 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.919298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.919323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.919351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.919372 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:00Z","lastTransitionTime":"2025-10-09T08:18:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.978968 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:18:00 crc kubenswrapper[4872]: I1009 08:18:00.980442 4872 scope.go:117] "RemoveContainer" containerID="2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1" Oct 09 08:18:00 crc kubenswrapper[4872]: E1009 08:18:00.980824 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.004834 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.022410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.023205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.023264 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.023300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.023323 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.027504 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.048772 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.061812 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.078005 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.093950 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.110032 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.122098 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.126411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.126462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.126478 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.126501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.126516 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.140520 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.155394 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.169489 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.180946 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.194043 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.208106 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.227593 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.228978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.229006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.229015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.229030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.229039 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.250781 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:01Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.331368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.331399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.331406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.331420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.331429 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.439116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.439196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.439235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.439269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.439293 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.460587 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.460663 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.460664 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:01 crc kubenswrapper[4872]: E1009 08:18:01.460805 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:01 crc kubenswrapper[4872]: E1009 08:18:01.460916 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:01 crc kubenswrapper[4872]: E1009 08:18:01.461025 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.542361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.542445 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.542457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.542479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.542491 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.646385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.646463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.646500 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.646519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.646535 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.750112 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.750154 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.750162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.750178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.750188 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.853738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.853813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.853838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.853888 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.853912 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.957107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.957213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.957246 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.957281 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:01 crc kubenswrapper[4872]: I1009 08:18:01.957305 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:01Z","lastTransitionTime":"2025-10-09T08:18:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.060426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.060484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.060498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.060521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.060534 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.163102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.163154 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.163164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.163185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.163197 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.266174 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.266273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.266298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.266339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.266367 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.369825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.369887 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.369904 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.369926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.369939 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.460772 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:02 crc kubenswrapper[4872]: E1009 08:18:02.460991 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.472586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.472659 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.472673 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.472691 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.472703 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.481500 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.499012 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.517510 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.535261 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.554708 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.575529 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.575568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.575581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.575600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.575622 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.577606 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.605442 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.620885 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.633333 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.647685 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.666449 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.682715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.682917 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.682980 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.683016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.683035 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.690802 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.703541 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.718058 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.729092 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.747828 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:02Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.786616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.786678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.786687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.786704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.786715 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.889306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.889360 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.889372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.889392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.889404 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.992701 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.992766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.992785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.992811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:02 crc kubenswrapper[4872]: I1009 08:18:02.992829 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:02Z","lastTransitionTime":"2025-10-09T08:18:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.096095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.096189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.096213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.096246 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.096276 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.200095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.200163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.200201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.200236 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.200257 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.303779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.303842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.303857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.303879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.303894 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.407169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.407240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.407254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.407274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.407323 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.460957 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.460977 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.461061 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:03 crc kubenswrapper[4872]: E1009 08:18:03.461185 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:03 crc kubenswrapper[4872]: E1009 08:18:03.461531 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:03 crc kubenswrapper[4872]: E1009 08:18:03.461563 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.516750 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:03 crc kubenswrapper[4872]: E1009 08:18:03.517030 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:03 crc kubenswrapper[4872]: E1009 08:18:03.517154 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:11.517125691 +0000 UTC m=+49.707654347 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.519604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.519634 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.519667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.519687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.519702 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.623271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.623330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.623342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.623364 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.623377 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.725922 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.725961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.725973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.725990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.726001 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.828624 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.828716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.828735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.828761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.828779 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.931437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.931498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.931520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.931546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:03 crc kubenswrapper[4872]: I1009 08:18:03.931565 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:03Z","lastTransitionTime":"2025-10-09T08:18:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.034927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.034984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.035001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.035026 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.035042 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.137610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.137668 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.137680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.137698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.137708 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.240814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.240865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.240876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.240893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.240903 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.343690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.343743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.343756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.343776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.343790 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.447027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.447107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.447130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.447163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.447186 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.461550 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:04 crc kubenswrapper[4872]: E1009 08:18:04.461846 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.550283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.550351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.550373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.550402 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.550421 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.654211 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.654289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.654313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.654338 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.654355 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.757822 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.757914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.757938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.757966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.757985 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.861170 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.861687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.861882 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.862106 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.862292 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.966037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.966398 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.966540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.966705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:04 crc kubenswrapper[4872]: I1009 08:18:04.966824 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:04Z","lastTransitionTime":"2025-10-09T08:18:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.069979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.070043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.070061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.070091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.070111 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.173955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.174025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.174043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.174068 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.174091 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.277185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.277254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.277273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.277298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.277317 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.380759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.380813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.380832 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.380862 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.380885 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.461159 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.461158 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:05 crc kubenswrapper[4872]: E1009 08:18:05.461386 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:05 crc kubenswrapper[4872]: E1009 08:18:05.461507 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.461172 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:05 crc kubenswrapper[4872]: E1009 08:18:05.461811 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.483789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.483829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.483844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.483862 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.483878 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.587193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.587254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.587267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.587286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.587298 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.690134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.690204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.690221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.690247 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.690268 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.793877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.793940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.793958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.793983 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.794001 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.896723 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.896769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.896785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.896809 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:05 crc kubenswrapper[4872]: I1009 08:18:05.896827 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:05Z","lastTransitionTime":"2025-10-09T08:18:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.000164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.000220 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.000233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.000253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.000267 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.102900 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.102979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.102999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.103031 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.103056 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.206501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.206573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.206591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.206619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.206665 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.309745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.309807 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.309824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.309850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.309876 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.413686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.413779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.413798 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.413838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.413865 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.460996 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:06 crc kubenswrapper[4872]: E1009 08:18:06.461203 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.517081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.517164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.517187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.517218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.517243 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.620815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.620880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.620900 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.620928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.620947 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.724069 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.724145 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.724167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.724198 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.724228 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.826962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.827029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.827051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.827075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.827093 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.930163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.930210 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.930221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.930238 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:06 crc kubenswrapper[4872]: I1009 08:18:06.930250 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:06Z","lastTransitionTime":"2025-10-09T08:18:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.032944 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.032994 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.033010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.033030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.033041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.135392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.135459 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.135470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.135488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.135499 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.238570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.238610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.238620 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.238661 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.238674 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.343971 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.344100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.344124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.344159 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.344196 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.447777 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.447884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.447895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.447916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.447929 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.461342 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:07 crc kubenswrapper[4872]: E1009 08:18:07.461502 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.461576 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:07 crc kubenswrapper[4872]: E1009 08:18:07.461668 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.461733 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:07 crc kubenswrapper[4872]: E1009 08:18:07.461886 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.550494 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.550536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.550545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.550561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.550570 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.653327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.653365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.653377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.653397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.653410 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.756047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.756101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.756117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.756139 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.756157 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.859608 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.859725 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.859749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.859777 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.859800 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.962542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.962594 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.962612 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.962666 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:07 crc kubenswrapper[4872]: I1009 08:18:07.962686 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:07Z","lastTransitionTime":"2025-10-09T08:18:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.066106 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.066186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.066221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.066255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.066275 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.169194 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.169262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.169297 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.169326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.169352 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.273026 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.273111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.273147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.273178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.273199 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.363891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.363966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.363978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.363997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.364008 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.386568 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:08Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.393725 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.393826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.393858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.393890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.393913 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.413910 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:08Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.419132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.419193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.419217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.419251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.419275 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.440706 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:08Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.447548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.447600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.447618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.447688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.447726 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.461333 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.461538 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.469965 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:08Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.474677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.474738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.474757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.474780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.474798 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.490365 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:08Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:08 crc kubenswrapper[4872]: E1009 08:18:08.490500 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.492255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.492290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.492301 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.492317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.492331 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.595573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.595679 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.595704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.595733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.595753 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.702076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.702146 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.702165 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.702191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.702209 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.805253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.805296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.805305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.805322 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.805333 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.908782 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.908849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.908864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.908888 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:08 crc kubenswrapper[4872]: I1009 08:18:08.908904 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:08Z","lastTransitionTime":"2025-10-09T08:18:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.012488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.012560 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.012578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.012604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.012622 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.115971 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.116008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.116017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.116033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.116041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.219455 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.219540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.219583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.219617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.219678 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.323231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.323292 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.323309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.323331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.323346 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.426404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.426476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.426489 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.426508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.426519 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.461342 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.461474 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:09 crc kubenswrapper[4872]: E1009 08:18:09.461510 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.461353 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:09 crc kubenswrapper[4872]: E1009 08:18:09.461603 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:09 crc kubenswrapper[4872]: E1009 08:18:09.461803 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.529321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.529393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.529411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.529440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.529462 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.633048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.633126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.633143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.633172 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.633190 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.735589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.735672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.735685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.735701 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.735714 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.838937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.839016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.839030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.839054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.839068 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.942256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.942316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.942330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.942350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:09 crc kubenswrapper[4872]: I1009 08:18:09.942363 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:09Z","lastTransitionTime":"2025-10-09T08:18:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.044986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.045055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.045074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.045102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.045121 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.148118 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.148185 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.148204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.148234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.148256 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.251367 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.251446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.251469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.251501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.251528 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.354589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.354695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.354719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.354760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.354782 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.457770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.457865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.457890 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.457921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.457943 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.461015 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:10 crc kubenswrapper[4872]: E1009 08:18:10.461184 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.561213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.561291 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.561303 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.561322 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.561333 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.663252 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.663309 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.663319 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.663336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.663346 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.766474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.766561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.766586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.766616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.766681 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.868677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.868720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.868731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.868747 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.868757 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.970939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.970984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.970993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.971008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:10 crc kubenswrapper[4872]: I1009 08:18:10.971036 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:10Z","lastTransitionTime":"2025-10-09T08:18:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.074330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.074393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.074410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.074436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.074459 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.177204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.177275 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.177290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.177311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.177328 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.279986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.280031 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.280041 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.280056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.280065 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.382827 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.382891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.382901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.382936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.382949 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.461000 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.461039 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:11 crc kubenswrapper[4872]: E1009 08:18:11.461189 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.461293 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:11 crc kubenswrapper[4872]: E1009 08:18:11.461442 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:11 crc kubenswrapper[4872]: E1009 08:18:11.461543 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.485049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.485100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.485116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.485136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.485149 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.587271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.587338 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.587354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.587379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.587396 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.602550 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.602634 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:11 crc kubenswrapper[4872]: E1009 08:18:11.602899 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:11 crc kubenswrapper[4872]: E1009 08:18:11.603009 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:27.602972529 +0000 UTC m=+65.793501195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.613476 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.618325 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.631679 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.643485 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.658842 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.673620 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.689868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.689919 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.689934 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.689960 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.689976 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.690929 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.705031 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.715606 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.729073 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.740483 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.753214 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.769593 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.784447 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.792490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.792528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.792537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.792551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.792561 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.794314 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.810666 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.830411 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:11Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.894786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.894828 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.894836 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.894853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:11 crc kubenswrapper[4872]: I1009 08:18:11.894863 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:11Z","lastTransitionTime":"2025-10-09T08:18:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.015954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.015999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.016007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.016023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.016032 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.118167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.118213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.118224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.118241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.118252 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.220752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.220804 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.220814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.220829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.220840 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.323820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.323866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.323877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.323896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.323907 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.427738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.427781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.427795 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.428966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.429044 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.461535 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:12 crc kubenswrapper[4872]: E1009 08:18:12.461749 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.475766 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.489551 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.504694 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.517082 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532140 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532152 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.532723 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.550122 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.567101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.581564 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.593004 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.605215 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.614800 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.625885 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.635011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.635070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.635080 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.635096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.635110 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.638829 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.649752 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.661719 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.675602 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.697334 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:12Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.736875 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.736914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.736925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.736940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.736949 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.839047 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.839091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.839102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.839119 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.839131 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.942390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.942470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.942498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.942530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:12 crc kubenswrapper[4872]: I1009 08:18:12.942554 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:12Z","lastTransitionTime":"2025-10-09T08:18:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.046030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.046076 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.046088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.046108 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.046122 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.150269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.150371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.150401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.150439 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.150473 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.225210 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.225401 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225436 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:18:45.225406708 +0000 UTC m=+83.415935374 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.225501 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.225562 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.225608 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225715 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225742 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225764 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225794 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225824 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225836 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:45.22581176 +0000 UTC m=+83.416340426 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225926 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:45.225902663 +0000 UTC m=+83.416431329 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225957 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:45.225940824 +0000 UTC m=+83.416469490 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225957 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.225998 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.226017 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.226101 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:45.226074028 +0000 UTC m=+83.416602694 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.253808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.253860 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.253876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.253901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.253918 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.356805 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.356880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.356904 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.356937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.356960 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460408 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460724 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460767 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.460820 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.460870 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.461076 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:13 crc kubenswrapper[4872]: E1009 08:18:13.461230 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.563543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.563602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.563613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.563688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.563701 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.666428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.666493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.666512 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.666542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.666565 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.769210 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.769261 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.769273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.769294 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.769307 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.871390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.871454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.871475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.871499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.871514 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.974585 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.974686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.974699 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.974718 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:13 crc kubenswrapper[4872]: I1009 08:18:13.974733 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:13Z","lastTransitionTime":"2025-10-09T08:18:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.077591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.077688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.077706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.077731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.077748 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.180687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.180753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.180770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.180791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.180807 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.283420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.283463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.283473 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.283490 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.283503 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.386025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.386085 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.386099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.386117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.386128 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.461420 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:14 crc kubenswrapper[4872]: E1009 08:18:14.461668 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.488002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.488045 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.488055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.488069 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.488083 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.590506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.590546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.590556 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.590573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.590585 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.693099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.693135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.693144 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.693178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.693196 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.796324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.796378 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.796389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.796409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.796422 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.899728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.899778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.899790 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.899813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:14 crc kubenswrapper[4872]: I1009 08:18:14.899827 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:14Z","lastTransitionTime":"2025-10-09T08:18:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.001903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.001949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.001959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.001975 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.001987 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.104605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.104722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.104748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.104777 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.104800 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.207910 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.207962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.207977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.207995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.208006 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.310961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.311023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.311040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.311064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.311081 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.413702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.413762 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.413774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.413792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.413804 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.461467 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.461468 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.461494 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:15 crc kubenswrapper[4872]: E1009 08:18:15.461818 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:15 crc kubenswrapper[4872]: E1009 08:18:15.462054 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:15 crc kubenswrapper[4872]: E1009 08:18:15.462312 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.463866 4872 scope.go:117] "RemoveContainer" containerID="2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.516603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.517088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.517104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.517126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.517139 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.620228 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.620265 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.620275 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.620290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.620300 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.723659 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.723697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.723707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.723721 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.723730 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.826946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.826997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.827009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.827030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.827041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.835452 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/1.log" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.838108 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.838475 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.855442 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.876376 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.890518 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.902903 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.916602 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.926533 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.929013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.929046 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.929056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.929075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.929089 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:15Z","lastTransitionTime":"2025-10-09T08:18:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.940447 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:15 crc kubenswrapper[4872]: I1009 08:18:15.980925 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.001064 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:15Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.015118 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.026825 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.031415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.031462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.031474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.031492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.031504 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.039600 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.053305 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.066919 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.079926 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.090507 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.104257 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.133913 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.133963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.133973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.133990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.134000 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.237600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.237663 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.237674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.237692 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.237703 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.340475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.340522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.340534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.340557 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.340570 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.444410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.444466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.444478 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.444502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.444513 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.461599 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:16 crc kubenswrapper[4872]: E1009 08:18:16.461760 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.547902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.547953 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.547964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.547982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.547993 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.650256 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.650331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.650354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.650384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.650406 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.754123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.754179 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.754192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.754214 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.754226 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.846884 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/2.log" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.848309 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/1.log" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.851928 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" exitCode=1 Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.852306 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.852466 4872 scope.go:117] "RemoveContainer" containerID="2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.853273 4872 scope.go:117] "RemoveContainer" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" Oct 09 08:18:16 crc kubenswrapper[4872]: E1009 08:18:16.853806 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.859913 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.859983 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.860003 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.860040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.860060 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.876801 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.892756 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.908117 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.922968 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.937831 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.949197 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.963457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.963499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.963510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.963527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.963539 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:16Z","lastTransitionTime":"2025-10-09T08:18:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.965279 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.983431 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2219dfbe2204c0a39596cd840d04d78589b35fbcfa2b0ea094d58ad0f0c4bbc1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"message\\\":\\\":53.534813 6314 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.534904 6314 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535106 6314 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:17:53.535385 6314 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:17:53.535779 6314 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1009 08:17:53.535847 6314 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1009 08:17:53.535856 6314 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1009 08:17:53.535896 6314 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1009 08:17:53.535897 6314 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1009 08:17:53.535913 6314 handler.go:208] Removed *v1.Node event handler 2\\\\nI1009 08:17:53.535950 6314 factory.go:656] Stopping watch factory\\\\nI1009 08:17:53.535950 6314 handler.go:208] Removed *v1.Node event handler 7\\\\nI1009 08:17:53.535972 6314 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:17:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:52Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:16 crc kubenswrapper[4872]: I1009 08:18:16.995433 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:16Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.007118 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.017698 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.027408 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.039226 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.053850 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.065863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.065913 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.065924 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.065940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.065954 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.066697 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.079289 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.088385 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.168753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.168793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.168801 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.168820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.168829 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.271355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.271414 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.271434 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.271458 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.271474 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.374263 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.374331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.374346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.374366 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.374382 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.460921 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.460940 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.461039 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:17 crc kubenswrapper[4872]: E1009 08:18:17.461248 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:17 crc kubenswrapper[4872]: E1009 08:18:17.461419 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:17 crc kubenswrapper[4872]: E1009 08:18:17.461505 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.477293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.477335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.477345 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.477362 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.477373 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.580097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.580188 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.580215 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.580254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.580280 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.683209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.683248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.683258 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.683278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.683289 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.786029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.786075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.786084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.786099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.786111 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.857805 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/2.log" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.862976 4872 scope.go:117] "RemoveContainer" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" Oct 09 08:18:17 crc kubenswrapper[4872]: E1009 08:18:17.863268 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.880192 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.888362 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.888422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.888436 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.888454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.888467 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.896359 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.917778 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.952129 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.972816 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.988872 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:17Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.990932 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.990974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.990986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.991006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:17 crc kubenswrapper[4872]: I1009 08:18:17.991018 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:17Z","lastTransitionTime":"2025-10-09T08:18:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.006622 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.018843 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.036856 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.054592 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.069311 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.082009 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.094046 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.097810 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.097842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.097852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.097870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.097881 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.111092 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.125773 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.139250 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.154341 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.201285 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.201357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.201379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.201410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.201431 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.304508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.304634 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.304692 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.304715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.304739 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.407544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.407591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.407603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.407625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.407659 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.461420 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.461612 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.496913 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.496951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.497005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.497032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.497050 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.513704 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.518212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.518248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.518260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.518307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.518318 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.531012 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.535211 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.535250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.535263 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.535283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.535297 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.546710 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.550517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.550558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.550568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.550606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.550618 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.562600 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.566741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.566802 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.566823 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.566853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.566875 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.585538 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:18Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:18 crc kubenswrapper[4872]: E1009 08:18:18.585908 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.588134 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.588167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.588175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.588189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.588200 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.690578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.690686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.690712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.690742 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.690766 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.793966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.794019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.794031 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.794049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.794059 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.896484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.896527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.896540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.896557 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.896567 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.998760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.998808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.998819 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.998837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:18 crc kubenswrapper[4872]: I1009 08:18:18.998849 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:18Z","lastTransitionTime":"2025-10-09T08:18:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.101820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.102195 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.102341 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.102488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.102623 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.206164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.206593 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.206784 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.206936 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.207068 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.309378 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.309444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.309457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.309476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.309490 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.412450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.412534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.412548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.412569 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.412581 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.461163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.461228 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.461195 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:19 crc kubenswrapper[4872]: E1009 08:18:19.461390 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:19 crc kubenswrapper[4872]: E1009 08:18:19.461461 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:19 crc kubenswrapper[4872]: E1009 08:18:19.461596 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.515685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.515725 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.515737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.515754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.515767 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.618426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.618476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.618488 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.618506 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.618516 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.721807 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.721878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.721896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.721918 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.721936 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.825371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.825482 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.825502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.825526 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.825545 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.928759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.928847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.928868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.928908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:19 crc kubenswrapper[4872]: I1009 08:18:19.928951 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:19Z","lastTransitionTime":"2025-10-09T08:18:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.032808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.032873 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.032889 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.032914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.032932 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.135457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.135523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.135547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.135570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.135584 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.237945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.237994 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.238012 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.238035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.238053 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.341019 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.341084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.341104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.341129 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.341150 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.444164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.444286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.444313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.444347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.444370 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.461559 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:20 crc kubenswrapper[4872]: E1009 08:18:20.461855 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.547469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.547577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.547589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.547611 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.547627 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.650961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.651059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.651081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.651099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.651110 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.753839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.753916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.753926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.753941 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.753951 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.857685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.857748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.857765 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.857789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.857809 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.960516 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.960630 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.960903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.960925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:20 crc kubenswrapper[4872]: I1009 08:18:20.960938 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:20Z","lastTransitionTime":"2025-10-09T08:18:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.063003 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.063795 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.063830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.063855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.063869 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.166926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.166983 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.167001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.167023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.167038 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.270466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.270538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.270561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.270590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.270615 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.373393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.373446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.373471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.373487 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.373497 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.460895 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.460949 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.460989 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:21 crc kubenswrapper[4872]: E1009 08:18:21.461068 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:21 crc kubenswrapper[4872]: E1009 08:18:21.461210 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:21 crc kubenswrapper[4872]: E1009 08:18:21.461386 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.476857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.476902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.476914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.476933 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.476946 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.579991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.580048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.580066 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.580090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.580107 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.682695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.682753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.682770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.682794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.682811 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.785268 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.785315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.785327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.785346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.785358 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.888037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.888073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.888081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.888096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.888105 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.990775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.990822 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.990834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.990848 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:21 crc kubenswrapper[4872]: I1009 08:18:21.990858 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:21Z","lastTransitionTime":"2025-10-09T08:18:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.093633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.093688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.093700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.093716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.093729 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.196896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.196966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.196981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.197006 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.197060 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.300605 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.300683 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.300695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.300733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.300745 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.403156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.403206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.403220 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.403239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.403250 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.460593 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:22 crc kubenswrapper[4872]: E1009 08:18:22.460770 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.477206 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.490952 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.504634 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.507012 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.507045 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.507075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.507200 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.507249 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.517019 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.538804 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.565102 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.580288 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.593555 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.605088 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.609554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.609592 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.609602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.609619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.609630 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.615694 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.627499 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.638576 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.652695 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.665974 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.675533 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.688207 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.701751 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:22Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.711782 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.711838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.711848 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.711867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.711882 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.813677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.813720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.813731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.813752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.813769 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.916246 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.916293 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.916304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.916323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:22 crc kubenswrapper[4872]: I1009 08:18:22.916335 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:22Z","lastTransitionTime":"2025-10-09T08:18:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.018626 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.018682 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.018691 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.018709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.018721 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.120789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.120840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.120854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.120873 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.120886 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.224207 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.224271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.224282 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.224301 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.224313 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.327467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.327519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.327532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.327551 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.327563 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.430332 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.430410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.430426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.430451 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.430468 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.461121 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.461292 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:23 crc kubenswrapper[4872]: E1009 08:18:23.461363 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.461453 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:23 crc kubenswrapper[4872]: E1009 08:18:23.461528 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:23 crc kubenswrapper[4872]: E1009 08:18:23.461708 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.533298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.533534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.533547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.533566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.533581 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.636833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.636881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.636891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.636909 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.636921 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.740082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.740122 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.740132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.740149 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.740159 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.842042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.842083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.842097 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.842116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.842129 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.945208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.945285 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.945296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.945317 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:23 crc kubenswrapper[4872]: I1009 08:18:23.945330 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:23Z","lastTransitionTime":"2025-10-09T08:18:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.047770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.047857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.047870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.047896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.047908 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.150839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.150876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.150884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.150899 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.150909 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.254413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.254510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.254571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.254599 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.254634 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.357328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.357389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.357401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.357415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.357425 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460670 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460683 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460692 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.460693 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: E1009 08:18:24.460825 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.563471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.563538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.563558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.563584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.563603 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.666754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.666808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.666819 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.666839 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.666850 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.769956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.770004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.770013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.770029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.770040 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.873699 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.873776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.873787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.873806 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.873819 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.976837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.976907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.976921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.976940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:24 crc kubenswrapper[4872]: I1009 08:18:24.976953 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:24Z","lastTransitionTime":"2025-10-09T08:18:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.079946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.079990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.080002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.080021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.080032 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.182165 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.182205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.182235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.182251 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.182261 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.285300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.285330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.285338 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.285352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.285361 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.387816 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.387845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.387853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.387868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.387877 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.460775 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:25 crc kubenswrapper[4872]: E1009 08:18:25.460954 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.461382 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:25 crc kubenswrapper[4872]: E1009 08:18:25.461452 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.461492 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:25 crc kubenswrapper[4872]: E1009 08:18:25.461551 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.490681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.490727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.490749 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.490770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.490782 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.593091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.593133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.593143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.593156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.593167 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.696064 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.696100 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.696109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.696125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.696134 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.799536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.799618 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.799729 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.799789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.799803 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.901905 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.901959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.901977 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.901996 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:25 crc kubenswrapper[4872]: I1009 08:18:25.902006 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:25Z","lastTransitionTime":"2025-10-09T08:18:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.004972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.005030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.005043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.005058 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.005068 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.108108 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.108157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.108171 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.108199 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.108210 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.210950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.210992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.211002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.211041 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.211056 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.313924 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.313978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.313997 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.314024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.314041 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.416714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.416753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.416763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.416778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.416789 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.461319 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:26 crc kubenswrapper[4872]: E1009 08:18:26.461541 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.518927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.518995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.519009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.519029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.519042 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.621243 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.621287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.621297 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.621313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.621325 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.723998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.724043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.724056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.724075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.724088 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.827796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.827855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.827868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.827892 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.827905 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.930416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.930858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.931043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.931263 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:26 crc kubenswrapper[4872]: I1009 08:18:26.931407 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:26Z","lastTransitionTime":"2025-10-09T08:18:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.034600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.034687 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.034706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.034730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.034747 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.137856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.137907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.137921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.137943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.137955 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.241518 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.241556 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.241568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.241585 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.241597 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.344702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.344741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.344756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.344773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.344783 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.447787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.447846 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.447858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.447879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.447892 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.461179 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.461249 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:27 crc kubenswrapper[4872]: E1009 08:18:27.461345 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:27 crc kubenswrapper[4872]: E1009 08:18:27.461530 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.461740 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:27 crc kubenswrapper[4872]: E1009 08:18:27.462014 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.550011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.550060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.550073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.550091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.550102 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.652730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.652778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.652789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.652806 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.652836 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.673548 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:27 crc kubenswrapper[4872]: E1009 08:18:27.673783 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:27 crc kubenswrapper[4872]: E1009 08:18:27.674108 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:18:59.6740859 +0000 UTC m=+97.864614526 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.755796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.755876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.755896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.755955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.755971 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.858996 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.859042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.859054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.859074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.859086 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.961902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.961973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.961987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.962009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:27 crc kubenswrapper[4872]: I1009 08:18:27.962025 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:27Z","lastTransitionTime":"2025-10-09T08:18:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.064777 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.064841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.064852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.064869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.064879 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.166475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.166540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.166559 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.166578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.166589 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.269142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.269181 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.269193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.269213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.269225 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.371702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.371766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.371780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.371818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.371839 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.461145 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.461425 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.473706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.473748 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.473758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.473773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.473786 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.576362 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.576408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.576420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.576437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.576448 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.640353 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.640393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.640404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.640422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.640464 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.653568 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:28Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.657981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.658005 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.658013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.658027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.658037 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.673074 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:28Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.677190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.677226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.677234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.677249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.677261 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.692058 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:28Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.696209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.696252 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.696268 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.696291 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.696306 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.708459 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:28Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.712033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.712089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.712101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.712120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.712130 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.725424 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:28Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:28Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:28 crc kubenswrapper[4872]: E1009 08:18:28.725559 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.726961 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.727001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.727016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.727035 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.727047 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.829253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.829363 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.829376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.829395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.829409 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.932851 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.932891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.932902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.932919 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:28 crc kubenswrapper[4872]: I1009 08:18:28.932929 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:28Z","lastTransitionTime":"2025-10-09T08:18:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.035065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.035114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.035124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.035144 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.035155 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.137457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.137501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.137510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.137523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.137532 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.240102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.240187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.240208 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.240236 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.240256 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.343569 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.343627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.343661 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.343684 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.343697 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.446516 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.446665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.446689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.446716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.446735 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.460864 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.460915 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.460870 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:29 crc kubenswrapper[4872]: E1009 08:18:29.461053 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:29 crc kubenswrapper[4872]: E1009 08:18:29.461133 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:29 crc kubenswrapper[4872]: E1009 08:18:29.461305 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.549276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.549314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.549332 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.549350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.549365 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.651763 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.651793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.651803 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.651821 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.651831 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.754060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.754101 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.754111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.754125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.754134 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.856720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.856793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.856808 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.856827 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.856840 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.903182 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/0.log" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.903232 4872 generic.go:334] "Generic (PLEG): container finished" podID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" containerID="956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90" exitCode=1 Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.903258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerDied","Data":"956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.903589 4872 scope.go:117] "RemoveContainer" containerID="956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.918282 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.931229 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.943573 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.955316 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.959192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.959219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.959228 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.959241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.959250 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:29Z","lastTransitionTime":"2025-10-09T08:18:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.971370 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.984658 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:29 crc kubenswrapper[4872]: I1009 08:18:29.998001 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:29Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.010275 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.024326 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.036920 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.048343 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.063410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.063461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.063474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.063491 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.063521 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.065852 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.089863 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.105319 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.119551 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.140101 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.153816 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.165841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.165884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.165893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.165909 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.165920 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.268728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.268769 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.268781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.268801 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.268812 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.371467 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.371498 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.371507 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.371522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.371531 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.461229 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:30 crc kubenswrapper[4872]: E1009 08:18:30.461390 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.474010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.474054 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.474065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.474083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.474093 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.577627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.577691 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.577700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.577715 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.577724 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.680168 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.680205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.680216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.680233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.680256 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.782950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.782991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.783001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.783017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.783027 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.886825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.886893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.886902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.886917 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.886929 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.909971 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/0.log" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.910058 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerStarted","Data":"7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc"} Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.931859 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.946973 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.959882 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.972909 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.989193 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:30Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.990682 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.990705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.990716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.990760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:30 crc kubenswrapper[4872]: I1009 08:18:30.990773 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:30Z","lastTransitionTime":"2025-10-09T08:18:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.005534 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.022694 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.036687 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.054968 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.070958 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.085185 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.093657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.093700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.093712 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.093731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.093746 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.098799 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.110858 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.123815 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.133600 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.148837 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.168158 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:31Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.196114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.196152 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.196164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.196180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.196191 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.298776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.298830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.298844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.298866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.298880 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.400863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.400909 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.400921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.400937 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.400947 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.460860 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.460918 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.460979 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:31 crc kubenswrapper[4872]: E1009 08:18:31.461024 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:31 crc kubenswrapper[4872]: E1009 08:18:31.461130 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:31 crc kubenswrapper[4872]: E1009 08:18:31.461217 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.502938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.502979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.502991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.503008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.503020 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.605416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.605452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.605461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.605476 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.605487 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.707603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.707665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.707674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.707690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.707699 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.810982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.811104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.811132 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.811155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.811166 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.914169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.914217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.914230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.914245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:31 crc kubenswrapper[4872]: I1009 08:18:31.914254 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:31Z","lastTransitionTime":"2025-10-09T08:18:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.017052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.017107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.017120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.017139 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.017150 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.119840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.119876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.119886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.119900 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.119910 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.222444 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.222492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.222501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.222520 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.222532 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.328415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.328981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.328992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.329007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.329015 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.431815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.431868 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.431880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.431900 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.431911 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.461624 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:32 crc kubenswrapper[4872]: E1009 08:18:32.462089 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.462269 4872 scope.go:117] "RemoveContainer" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" Oct 09 08:18:32 crc kubenswrapper[4872]: E1009 08:18:32.462416 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.473585 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.484841 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.496458 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.505981 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.515778 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.524674 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.535331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.535384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.535405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.535428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.535441 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.537017 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.550835 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.566774 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.581435 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.597867 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.610016 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.626588 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.637901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.637926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.637935 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.637948 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.637957 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.645402 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.658257 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.676708 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.770789 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.771018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.771032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.771052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.771064 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.772421 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:32Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.873588 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.873624 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.873658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.873678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.873697 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.975857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.976095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.976191 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.976284 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:32 crc kubenswrapper[4872]: I1009 08:18:32.976414 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:32Z","lastTransitionTime":"2025-10-09T08:18:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.081782 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.081825 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.081836 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.081856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.081867 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.184745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.184800 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.184814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.184834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.184850 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.287720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.287764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.287775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.287794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.287806 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.390959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.391015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.391027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.391046 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.391058 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.460695 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.460695 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:33 crc kubenswrapper[4872]: E1009 08:18:33.461426 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.460854 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:33 crc kubenswrapper[4872]: E1009 08:18:33.461481 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:33 crc kubenswrapper[4872]: E1009 08:18:33.461720 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.494461 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.494523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.494536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.494558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.494573 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.598322 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.598379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.598391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.598409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.598419 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.701470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.701530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.701547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.701566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.701577 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.804546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.804591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.804600 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.804617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.804626 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.907406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.907485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.907501 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.907527 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:33 crc kubenswrapper[4872]: I1009 08:18:33.907541 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:33Z","lastTransitionTime":"2025-10-09T08:18:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.010211 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.010249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.010260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.010278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.010289 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.113055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.113123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.113131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.113151 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.113161 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.215227 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.215271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.215280 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.215298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.215310 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.317722 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.317757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.317766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.317781 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.317792 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.420625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.420674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.420751 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.420772 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.420783 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.462017 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:34 crc kubenswrapper[4872]: E1009 08:18:34.462195 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.523356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.523402 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.523412 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.523429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.523440 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.627335 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.627383 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.627394 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.627414 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.627426 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.729675 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.729727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.729737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.729753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.729762 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.832240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.832337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.832352 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.832371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.832383 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.934505 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.934550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.934562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.934584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:34 crc kubenswrapper[4872]: I1009 08:18:34.934596 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:34Z","lastTransitionTime":"2025-10-09T08:18:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.037139 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.037184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.037195 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.037213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.037225 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.139462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.139499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.139510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.139525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.139534 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.242356 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.242689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.242874 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.242955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.243023 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.345566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.345657 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.345671 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.345690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.345700 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.448161 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.448269 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.448279 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.448295 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.448305 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.460958 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:35 crc kubenswrapper[4872]: E1009 08:18:35.461051 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.461131 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:35 crc kubenswrapper[4872]: E1009 08:18:35.461324 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.461179 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:35 crc kubenswrapper[4872]: E1009 08:18:35.461723 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.549964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.549999 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.550009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.550025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.550035 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.653014 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.653082 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.653095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.653115 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.653133 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.755562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.755601 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.755610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.755627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.755656 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.858248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.858288 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.858304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.858326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.858338 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.960916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.960958 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.960968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.960982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:35 crc kubenswrapper[4872]: I1009 08:18:35.960991 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:35Z","lastTransitionTime":"2025-10-09T08:18:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.063598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.063681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.063695 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.063714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.063728 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.165928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.165969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.165981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.165998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.166010 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.268989 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.269030 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.269043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.269067 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.269079 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.371336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.371385 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.371393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.371410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.371421 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.460819 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:36 crc kubenswrapper[4872]: E1009 08:18:36.461056 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.473375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.473428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.473438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.473453 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.473467 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.576978 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.577022 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.577033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.577051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.577065 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.679672 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.679708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.679719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.679735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.679746 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.782756 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.782818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.782838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.782863 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.782880 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.885416 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.885466 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.885477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.885495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.885507 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.988597 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.988696 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.988707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.988727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:36 crc kubenswrapper[4872]: I1009 08:18:36.988759 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:36Z","lastTransitionTime":"2025-10-09T08:18:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.091094 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.091155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.091171 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.091190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.091202 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.194651 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.194737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.194903 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.194930 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.194948 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.298173 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.298442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.298519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.298584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.298667 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.401535 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.401568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.401580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.401593 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.401602 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.461224 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.461272 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:37 crc kubenswrapper[4872]: E1009 08:18:37.461521 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.461276 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:37 crc kubenswrapper[4872]: E1009 08:18:37.461738 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:37 crc kubenswrapper[4872]: E1009 08:18:37.461891 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.504292 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.504369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.504381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.504396 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.504407 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.606622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.606719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.606737 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.606757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.606772 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.711295 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.711361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.711372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.711388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.711402 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.814583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.814665 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.814676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.814694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.814704 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.929681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.929729 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.929742 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.929761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:37 crc kubenswrapper[4872]: I1009 08:18:37.929775 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:37Z","lastTransitionTime":"2025-10-09T08:18:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.032706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.032746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.032758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.032774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.032784 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.137429 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.137483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.137499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.137523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.137539 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.240497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.240561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.240576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.240597 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.240613 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.343689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.344042 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.344120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.344209 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.344299 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.446589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.446689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.446708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.446732 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.446750 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.461188 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:38 crc kubenswrapper[4872]: E1009 08:18:38.461318 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.549974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.550460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.550591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.550778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.550884 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.653196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.653248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.653267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.653290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.653309 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.757040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.757074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.757109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.757127 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.757138 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.860797 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.860865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.860884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.860908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.860925 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.964224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.964312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.964339 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.964372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:38 crc kubenswrapper[4872]: I1009 08:18:38.964396 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:38Z","lastTransitionTime":"2025-10-09T08:18:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.068025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.068849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.068901 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.069133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.069156 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.098216 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.098591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.099078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.099485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.100006 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.125356 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:39Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.130998 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.131135 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.131217 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.131299 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.131381 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.147600 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:39Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.153440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.153502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.153525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.153554 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.153572 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.170439 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:39Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.175175 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.175221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.175238 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.175259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.175274 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.192508 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:39Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.196050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.196103 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.196128 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.196158 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.196178 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.213296 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:39Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.213408 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.215131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.215188 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.215201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.215218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.215231 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.318290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.318327 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.318337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.318358 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.318369 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.421758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.421810 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.421821 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.421836 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.421845 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.461112 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.461151 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.461151 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.461304 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.461389 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:39 crc kubenswrapper[4872]: E1009 08:18:39.461459 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.524951 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.525000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.525011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.525029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.525042 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.633206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.633276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.633296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.633320 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.633340 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.736276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.736333 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.736349 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.736370 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.736384 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.839986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.840041 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.840051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.840071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.840083 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.943613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.943676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.943686 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.943706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:39 crc kubenswrapper[4872]: I1009 08:18:39.943716 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:39Z","lastTransitionTime":"2025-10-09T08:18:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.046940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.046979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.046990 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.047010 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.047024 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.150109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.150153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.150164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.150178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.150187 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.253328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.253393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.253406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.253422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.253434 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.356024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.356125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.356138 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.356162 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.356176 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.459301 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.459355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.459369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.459391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.459405 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.460907 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:40 crc kubenswrapper[4872]: E1009 08:18:40.461059 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.563211 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.563300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.563321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.563351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.563371 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.665942 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.665988 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.666011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.666036 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.666052 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.767550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.767587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.767596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.767611 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.767623 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.870033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.870084 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.870098 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.870117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.870129 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.972043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.972515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.972584 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.972691 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:40 crc kubenswrapper[4872]: I1009 08:18:40.972814 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:40Z","lastTransitionTime":"2025-10-09T08:18:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.076382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.076811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.076914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.077015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.077115 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.180442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.180530 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.180550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.180583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.180609 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.283337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.283395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.283403 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.283420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.283435 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.385437 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.385517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.385528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.385549 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.385561 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.461445 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.461532 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:41 crc kubenswrapper[4872]: E1009 08:18:41.461629 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.461771 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:41 crc kubenswrapper[4872]: E1009 08:18:41.461998 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:41 crc kubenswrapper[4872]: E1009 08:18:41.462144 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.488083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.488147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.488157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.488184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.488208 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.591114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.591535 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.591721 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.591831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.591915 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.694708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.694746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.694758 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.694775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.694788 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.798291 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.798330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.798368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.798389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.798403 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.901586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.901969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.902143 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.902290 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:41 crc kubenswrapper[4872]: I1009 08:18:41.902398 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:41Z","lastTransitionTime":"2025-10-09T08:18:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.005254 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.005358 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.005376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.005395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.005406 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.108411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.108463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.108479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.108502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.108518 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.211878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.211920 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.211934 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.211956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.211970 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.314909 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.314947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.314960 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.314976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.314987 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.417052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.417091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.417102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.417117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.417129 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.461566 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:42 crc kubenswrapper[4872]: E1009 08:18:42.461807 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.479250 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.494418 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.507564 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.518682 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.519438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.519464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.519472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.519484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.519494 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.533449 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.544527 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.560105 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.573109 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.590261 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.603829 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.620780 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.622233 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.622259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.622267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.622281 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.622290 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.634369 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.645805 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.656928 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.666558 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.681409 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.700180 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:42Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.725189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.725253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.725266 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.725284 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.725296 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.828454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.828509 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.828528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.828552 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.828569 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.930312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.930355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.930372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.930395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:42 crc kubenswrapper[4872]: I1009 08:18:42.930411 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:42Z","lastTransitionTime":"2025-10-09T08:18:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.032987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.033040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.033055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.033073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.033083 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.135160 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.135210 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.135222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.135239 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.135251 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.238241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.238283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.238292 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.238307 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.238316 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.341591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.341667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.341680 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.341697 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.341709 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.444008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.444295 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.444381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.444462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.444546 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.461096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.461185 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.461096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:43 crc kubenswrapper[4872]: E1009 08:18:43.461247 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:43 crc kubenswrapper[4872]: E1009 08:18:43.461316 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:43 crc kubenswrapper[4872]: E1009 08:18:43.461397 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.547308 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.547591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.547700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.547791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.547866 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.651008 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.651062 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.651077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.651095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.651109 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.754300 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.754615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.754764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.754849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.754962 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.857274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.857306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.857315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.857331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.857342 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.959264 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.959316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.959329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.959347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:43 crc kubenswrapper[4872]: I1009 08:18:43.959358 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:43Z","lastTransitionTime":"2025-10-09T08:18:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.061918 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.061960 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.061971 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.061985 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.061996 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.164700 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.164753 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.164778 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.164796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.164809 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.268623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.268720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.268738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.268765 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.268784 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.371562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.372048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.372241 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.372481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.372734 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.461047 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:44 crc kubenswrapper[4872]: E1009 08:18:44.461220 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.475420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.475465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.475480 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.475499 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.475513 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.577925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.577985 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.578002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.578029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.578046 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.682078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.682178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.682249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.682274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.682324 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.784770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.784819 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.784834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.784853 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.784872 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.887542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.887591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.887603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.887617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.887627 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.990315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.990357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.990366 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.990381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:44 crc kubenswrapper[4872]: I1009 08:18:44.990390 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:44Z","lastTransitionTime":"2025-10-09T08:18:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.093428 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.093493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.093510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.093535 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.093552 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.195938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.196029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.196050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.196073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.196122 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.300336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.300418 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.300446 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.300480 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.300502 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.317109 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317291 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.317253849 +0000 UTC m=+147.507782515 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.317394 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.317502 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.317570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317661 4872 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317746 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.317669 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317794 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317794 4872 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317823 4872 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317792 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317894 4872 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317766 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.317742894 +0000 UTC m=+147.508271550 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317912 4872 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.317963 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.31793572 +0000 UTC m=+147.508464376 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.318005 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.317990712 +0000 UTC m=+147.508519388 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.318031 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.318018152 +0000 UTC m=+147.508546878 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.403344 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.403382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.403391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.403405 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.403414 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.461031 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.461047 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.461128 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.461656 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.461766 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:45 crc kubenswrapper[4872]: E1009 08:18:45.461471 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.506433 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.506483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.506504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.506537 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.506561 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.609180 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.609240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.609259 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.609283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.609302 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.712734 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.712807 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.712831 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.712864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.712887 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.816164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.816201 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.816212 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.816230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.816241 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.919606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.919727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.919750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.919779 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:45 crc kubenswrapper[4872]: I1009 08:18:45.919801 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:45Z","lastTransitionTime":"2025-10-09T08:18:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.023577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.023690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.023708 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.023764 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.023781 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.126609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.126679 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.126692 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.126709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.126721 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.228907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.228957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.228969 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.228987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.229001 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.331199 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.331454 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.331469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.331519 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.331535 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.433798 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.433847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.433858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.433877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.433890 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.461528 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:46 crc kubenswrapper[4872]: E1009 08:18:46.461715 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.462771 4872 scope.go:117] "RemoveContainer" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.474187 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.537484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.537523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.537532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.537546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.537558 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.640280 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.640325 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.640336 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.640355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.640367 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.743615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.743714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.743734 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.743759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.743814 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.846517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.846556 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.846564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.846578 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.846587 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.949305 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.949372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.949386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.949408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.949423 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:46Z","lastTransitionTime":"2025-10-09T08:18:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.965992 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/2.log" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.968513 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.969746 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.983235 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:46 crc kubenswrapper[4872]: I1009 08:18:46.994981 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:46Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.006895 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.017189 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.030693 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.043099 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.051625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.051694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.051704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.051720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.051751 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.052323 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d49aee14-c565-44d4-9c87-c24bffa83703\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8ae3fb4ff28294fa996d1b34c20df9f9c146a1c8110f4fff78359fe37c68cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.070397 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.088572 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.105379 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.117723 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.126784 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.141601 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.153834 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.153870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.153879 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.153894 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.153902 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.159853 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.172705 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.185331 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.195077 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.205713 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.256481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.256534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.256544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.256558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.256568 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.359988 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.360364 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.360375 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.360409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.360422 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.461328 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.461423 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:47 crc kubenswrapper[4872]: E1009 08:18:47.461936 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.462144 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:47 crc kubenswrapper[4872]: E1009 08:18:47.462202 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:47 crc kubenswrapper[4872]: E1009 08:18:47.462326 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.463050 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.463081 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.463092 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.463106 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.463116 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.565759 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.565791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.565802 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.565817 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.565828 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.668780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.668829 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.668840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.668856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.668866 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.771169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.771423 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.771492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.771610 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.771709 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.874232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.874276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.874288 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.874306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.874324 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.973208 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/3.log" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.974080 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/2.log" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.977739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.977992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.978021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.978052 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.978073 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:47Z","lastTransitionTime":"2025-10-09T08:18:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.977763 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" exitCode=1 Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.977788 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.978431 4872 scope.go:117] "RemoveContainer" containerID="a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.978537 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:18:47 crc kubenswrapper[4872]: E1009 08:18:47.978721 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:47 crc kubenswrapper[4872]: I1009 08:18:47.995852 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.008799 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.023342 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.040484 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.052966 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.070521 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.080562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.080780 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.080865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.080952 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.081039 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.085048 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.098678 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.114134 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.127157 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.141303 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.155117 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.166414 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d49aee14-c565-44d4-9c87-c24bffa83703\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8ae3fb4ff28294fa996d1b34c20df9f9c146a1c8110f4fff78359fe37c68cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.177434 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.184059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.184105 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.184123 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.184150 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.184167 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.192209 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.216624 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a946c58f6d8c3df02fa12d4606ea35a5686461ed52f9030b49c4de47da535528\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:16Z\\\",\\\"message\\\":\\\"lient-go/informers/factory.go:160\\\\nI1009 08:18:16.243838 6580 reflector.go:311] Stopping reflector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1009 08:18:16.244091 6580 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1009 08:18:16.244364 6580 reflector.go:311] Stopping reflector *v1.AdminPolicyBasedExternalRoute (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/adminpolicybasedroute/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.244407 6580 factory.go:656] Stopping watch factory\\\\nI1009 08:18:16.244535 6580 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1009 08:18:16.255607 6580 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1009 08:18:16.255627 6580 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1009 08:18:16.255702 6580 ovnkube.go:599] Stopped ovnkube\\\\nI1009 08:18:16.255724 6580 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1009 08:18:16.255827 6580 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:47Z\\\",\\\"message\\\":\\\"ort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.245\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9192, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1009 08:18:47.278172 7009 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 08:18:47.278155 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-5ftzb\\\\nI1009 08:18:47.278181 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.231774 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.244945 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:48Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.286728 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.286793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.286805 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.286820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.286831 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.389155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.389200 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.389213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.389231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.389243 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.461498 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:48 crc kubenswrapper[4872]: E1009 08:18:48.461667 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.491702 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.491754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.491768 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.491786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.491800 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.594253 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.594312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.594329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.594354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.594371 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.697973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.698011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.698024 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.698040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.698050 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.800407 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.800459 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.800479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.800503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.800522 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.903007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.903058 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.903070 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.903089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.903101 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:48Z","lastTransitionTime":"2025-10-09T08:18:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.985515 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/3.log" Oct 09 08:18:48 crc kubenswrapper[4872]: I1009 08:18:48.990720 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:18:48 crc kubenswrapper[4872]: E1009 08:18:48.991024 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.007689 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.007743 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.007754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.007774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.007787 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.009321 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.034402 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:47Z\\\",\\\"message\\\":\\\"ort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.245\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9192, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1009 08:18:47.278172 7009 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 08:18:47.278155 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-5ftzb\\\\nI1009 08:18:47.278181 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.048437 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.061213 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.071725 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.082937 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.098133 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.109893 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.111920 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.111955 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.111964 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.111982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.111994 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.120863 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.132901 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.142435 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.154005 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.166474 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.177759 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.187704 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.199476 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.209393 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d49aee14-c565-44d4-9c87-c24bffa83703\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8ae3fb4ff28294fa996d1b34c20df9f9c146a1c8110f4fff78359fe37c68cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.213939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.214002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.214015 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.214033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.214045 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.221191 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.240535 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.240688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.240783 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.240869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.240942 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.253657 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.257558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.257599 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.257611 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.257628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.257657 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.269895 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.274032 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.274066 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.274099 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.274117 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.274128 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.285927 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.288993 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.289016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.289027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.289041 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.289051 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.300074 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.303496 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.303521 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.303548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.303563 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.303571 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.318898 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:49Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.319048 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.320806 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.320854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.320870 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.320895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.320914 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.423276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.423323 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.423337 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.423355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.423367 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.461027 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.461082 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.461107 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.461194 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.461364 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:49 crc kubenswrapper[4872]: E1009 08:18:49.461463 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.527523 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.527557 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.527567 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.527581 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.527591 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.630726 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.630786 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.630803 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.630824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.630840 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.733889 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.733962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.733986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.734016 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.734039 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.837835 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.837907 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.837933 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.837968 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.837992 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.940371 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.940408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.940419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.940435 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:49 crc kubenswrapper[4872]: I1009 08:18:49.940446 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:49Z","lastTransitionTime":"2025-10-09T08:18:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.043204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.043274 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.043296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.043324 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.043527 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.146168 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.146213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.146227 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.146244 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.146257 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.249073 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.249131 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.249147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.249166 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.249180 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.352840 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.353196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.353221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.353267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.353290 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.456235 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.456286 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.456298 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.456314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.456324 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.461332 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:50 crc kubenswrapper[4872]: E1009 08:18:50.461484 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.558738 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.558785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.558796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.558814 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.558824 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.661623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.661704 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.661716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.661736 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.661748 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.764891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.764950 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.764963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.764982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.764993 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.867349 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.867589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.867598 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.867613 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.867626 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.970374 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.970627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.970658 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.970677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:50 crc kubenswrapper[4872]: I1009 08:18:50.970689 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:50Z","lastTransitionTime":"2025-10-09T08:18:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.073576 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.073617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.073625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.073668 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.073680 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.176057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.176104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.176114 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.176130 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.176141 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.280003 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.280090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.280102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.280124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.280137 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.383886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.383943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.383956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.383975 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.383988 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.461245 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.461313 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.461245 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:51 crc kubenswrapper[4872]: E1009 08:18:51.461450 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:51 crc kubenswrapper[4872]: E1009 08:18:51.461620 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:51 crc kubenswrapper[4872]: E1009 08:18:51.461730 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.486986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.487017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.487027 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.487043 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.487053 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.589827 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.589869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.589877 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.589895 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.589909 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.693422 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.693481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.693502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.693528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.693548 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.796368 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.796456 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.796477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.796504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.796534 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.899566 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.899622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.899633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.899681 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:51 crc kubenswrapper[4872]: I1009 08:18:51.899694 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:51Z","lastTransitionTime":"2025-10-09T08:18:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.001943 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.001988 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.002002 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.002017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.002028 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.104884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.104935 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.104946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.104962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.104974 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.207735 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.207775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.207787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.207804 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.207816 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.310807 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.310855 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.310867 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.310884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.310901 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.412799 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.412861 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.412869 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.412899 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.412909 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.461209 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:52 crc kubenswrapper[4872]: E1009 08:18:52.461683 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.476370 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"18553ea5-5fb3-4436-ad96-f4b3c578b972\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e139e312bdae20cc24473cc4342c4a3ed9fc982d8362df0e4d8c9874008f776\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f59b40a742ff26e1a8a9e4c2cea8824a08d2ab7b8abacc8fe8b3841f395eb2ea\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://85dd007b8b8ad558018b136fbafb88cd86ab736cad2ba3a2de7b092ece26615b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://19bb29932d382451a9d3401d7a210484e3eb4a98cd9141ce293dddcad59b91b1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://cb0af58a2cc702ac06f379265f4ada5d20c06d5237121416fd02832359bcb712\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"pace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": dial tcp [::1]:6443: connect: connection refused\\\\nI1009 08:17:26.102310 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1009 08:17:26.103347 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-4173873539/tls.crt::/tmp/serving-cert-4173873539/tls.key\\\\\\\"\\\\nI1009 08:17:41.302924 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1009 08:17:41.307577 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1009 08:17:41.307611 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1009 08:17:41.307680 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1009 08:17:41.307702 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1009 08:17:41.323167 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1009 08:17:41.323220 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1009 08:17:41.323246 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1009 08:17:41.323255 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1009 08:17:41.323262 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1009 08:17:41.323270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1009 08:17:41.323633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1009 08:17:41.325709 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8356f6195fdce5c5ee1f970eaee0afa5c12acba5f588399203e8a93f3808a2fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d4936d922f50e5635b2541b26ad49a1eed8d94023e2e3f591f11e74fa1c17aca\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.491681 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.505534 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5cc7b6a1-93b1-455b-aa21-d8bb7574857e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://406c4f3e88c33a604e2b07450ce47009f7a32996a8de7cc56cd0d37dbc4f100e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ngxs4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-98kmz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.516091 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.516142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.516156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.516182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.516197 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.518879 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-8grb6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"326d4497-ac5f-4a64-afde-14ded51665b8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jczsj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:55Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-8grb6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.531521 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-bd2z9" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e46110a-3d1a-431d-8489-1af8359725a4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://acc99bbf1568933958015cd7af06844177add703f712697a7293bddc1bc980b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-7q8ss\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-bd2z9\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.546011 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5ftzb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5f349b9-7850-4b1b-86da-e89f7cfd32df\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:29Z\\\",\\\"message\\\":\\\"2025-10-09T08:17:44+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8\\\\n2025-10-09T08:17:44+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9a137525-61d9-474f-bf80-fc907c381ff8 to /host/opt/cni/bin/\\\\n2025-10-09T08:17:44Z [verbose] multus-daemon started\\\\n2025-10-09T08:17:44Z [verbose] Readiness Indicator file check\\\\n2025-10-09T08:18:29Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:18:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ssc5b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5ftzb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.568064 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://86a825a4e5f763390eb0075e95edcbdb3651a29c7a1ab58b26a13cd455e78f9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.586761 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.601660 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://496db4048fa8e3c5fc6a772a1ed36939cc3f9b2e18489694a0cc2c7e845a5016\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7f9e8c27498f7306b92b330b4bbefecf4ec25debcc430abde0cbf687073054d0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.613571 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1d947a2-2094-4762-bf2f-98bf802b2036\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2a7385767786d5cf4a967f709f2934a90c14672048e0ebb5e66530093336f740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd905aa3e255a3fdaa3f298fe93d9e267372f71c7840eb12913bcdcc2058785a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://565cb3005168c7f8d999260cc052d6beb10151afa87aeab447e34c71cbfe3b35\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3c2b019526af97e80d4eba76d1e81b13b241ce375b24bfcb4fca2bd78ac2eed\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.618966 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.619021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.619034 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.619057 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.619072 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.624209 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d49aee14-c565-44d4-9c87-c24bffa83703\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3f8ae3fb4ff28294fa996d1b34c20df9f9c146a1c8110f4fff78359fe37c68cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://555f1a14b13d492fc99b6290b515a55cc119dc9e69edfb53424268aacc532dac\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.636924 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6c37ec5288e6eb74c50b8e0973fbd031b2531d957689790b2e2a1050195ff85e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.649606 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"63b22313-f1d5-44b5-877b-e5aca8937437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:55Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32f18078acdab31f50a38ddae072369cb7493735028489d30bfbb11fdbec8e7f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0ceb7cf984e2105fb47d50883150ba237b3321710e724dc363bb27abf136259a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bvbxk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:54Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-s7ffs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.668718 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fac60b03-54bb-43e8-8994-3674d42b209b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-09T08:18:47Z\\\",\\\"message\\\":\\\"ort:false}, services.lbConfig{vips:[]string{\\\\\\\"10.217.5.245\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9192, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nF1009 08:18:47.278172 7009 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:47Z is after 2025-08-24T17:21:41Z]\\\\nI1009 08:18:47.278155 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-5ftzb\\\\nI1009 08:18:47.278181 7009 obj_retry.go:303] Retry object setup: *v1.Pod openshift-machine-config-opera\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-09T08:18:46Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85kdn\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-xnhdk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.680151 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"93730b2e-d415-4780-be74-3b7387dfa767\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1c8fdb4ed08011d2ef0c715f075252a15ccd9f6cc7ea33a907451f986205e71c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abf0bdcd67436b51d146f41a191c7abd839ff2dc339b2127ce565fe6f61ee3d8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec944b91ffca66d91d98c96e02c0a5f69a7b267fc44a94b9f9ec43fbe77e5129\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://974ae19cfabb65d40b27731a7ee8d3ba8c099e180ec847ad6709b4f001e77af7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:23Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:22Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.691044 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.700404 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-7f65m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4d45d7db-4c94-4b84-83b1-9049c7441fed\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5013176d09cb67efe6185ae01b45130d696e78e9c2369cec44f281648c36d76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ld2c5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:41Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-7f65m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.715734 4872 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7bc7b555-0389-4f9c-9bc9-02387eb2f7c3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-09T08:17:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2dcd4ea32364080d697dfc7ef621f079e54260a7925a1e3929d218c60a3811a8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://22b508ec574ec4c97ffa35758a4fa69ae5feecce3ab5002231e44f11df080e78\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b03daf38b28d5ea72cc753722a9f7f30227afaf809148b13afa2bdd12f7a588\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7dc4cd6f124ae46048e4bb624f64ddc724d2f21f2e2f69f1772dac79481ed983\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3a22411368011c335a9e7feb79cc2a0fb5e05ab74f3e085fd12b104efd52e060\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91b89f59bdc36c287ec47d9cd42d87bc6aeb0aa2077e743989ceaeee2110cb76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://115f22dc0af3768dd114eaa08f90aa41a73621c00797fa1bca4798f503a09daf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-09T08:17:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-09T08:17:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-snww9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-09T08:17:42Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-ssrfz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:52Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.722351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.722380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.722392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.722408 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.722418 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.824940 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.824982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.824992 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.825007 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.825016 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.928096 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.928164 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.928187 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.928218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:52 crc kubenswrapper[4872]: I1009 08:18:52.928241 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:52Z","lastTransitionTime":"2025-10-09T08:18:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.030590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.030676 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.030694 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.030720 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.030737 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.133443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.133510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.133558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.133575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.133586 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.235572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.235616 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.235626 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.235660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.235672 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.338477 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.338532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.338546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.338572 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.338585 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.441619 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.441710 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.441727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.441751 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.441767 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.461342 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.461363 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.461387 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:53 crc kubenswrapper[4872]: E1009 08:18:53.461501 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:53 crc kubenswrapper[4872]: E1009 08:18:53.461749 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:53 crc kubenswrapper[4872]: E1009 08:18:53.462093 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.544273 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.544314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.544330 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.544348 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.544358 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.647504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.647560 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.647583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.647604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.647618 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.750332 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.750376 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.750386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.750402 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.750412 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.853023 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.853102 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.853126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.853156 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.853178 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.956415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.956484 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.956508 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.956541 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:53 crc kubenswrapper[4872]: I1009 08:18:53.956565 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:53Z","lastTransitionTime":"2025-10-09T08:18:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.059481 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.059568 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.059603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.059633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.059702 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.162821 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.162884 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.162908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.162941 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.162965 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.266133 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.266214 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.266245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.266276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.266298 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.370095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.370161 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.370184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.370213 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.370232 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.460983 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:54 crc kubenswrapper[4872]: E1009 08:18:54.461182 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.473856 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.473914 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.473931 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.473948 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.473965 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.578287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.578350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.578367 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.578393 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.578411 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.680805 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.680866 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.680885 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.680908 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.680925 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.783306 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.783377 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.783391 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.783406 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.783419 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.885957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.885991 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.886001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.886014 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.886025 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.988736 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.988773 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.988785 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.988802 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:54 crc kubenswrapper[4872]: I1009 08:18:54.988814 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:54Z","lastTransitionTime":"2025-10-09T08:18:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.090771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.090837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.090861 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.090892 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.090914 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.193205 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.193240 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.193250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.193287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.193296 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.296979 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.297045 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.297055 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.297071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.297081 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.400633 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.400705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.400719 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.400739 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.400756 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.461444 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.461510 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:55 crc kubenswrapper[4872]: E1009 08:18:55.461611 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.461529 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:55 crc kubenswrapper[4872]: E1009 08:18:55.461819 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:55 crc kubenswrapper[4872]: E1009 08:18:55.461947 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.504426 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.504469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.504479 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.504493 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.504503 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.607031 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.607111 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.607121 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.607136 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.607148 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.709845 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.709909 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.709921 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.709938 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.709950 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.812745 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.812809 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.812824 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.812844 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.812858 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.915770 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.916077 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.916153 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.916226 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:55 crc kubenswrapper[4872]: I1009 08:18:55.916304 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:55Z","lastTransitionTime":"2025-10-09T08:18:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.019166 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.019225 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.019234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.019250 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.019259 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.122465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.122514 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.122525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.122543 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.122556 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.225757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.225805 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.225815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.225833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.225842 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.328911 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.328962 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.328972 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.328987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.328999 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.431602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.431881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.431959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.432037 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.432219 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.461304 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:56 crc kubenswrapper[4872]: E1009 08:18:56.461466 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.534925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.534974 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.534984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.535004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.535015 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.638000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.638041 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.638049 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.638063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.638072 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.740876 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.740918 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.740929 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.740945 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.740956 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.843351 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.843404 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.843419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.843438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.843453 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.946048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.946095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.946107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.946126 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:56 crc kubenswrapper[4872]: I1009 08:18:56.946137 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:56Z","lastTransitionTime":"2025-10-09T08:18:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.048511 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.048558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.048570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.048587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.048597 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.150474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.150517 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.150528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.150544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.150554 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.253232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.253268 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.253277 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.253291 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.253302 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.356380 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.356432 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.356443 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.356465 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.356475 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461036 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:57 crc kubenswrapper[4872]: E1009 08:18:57.461217 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461328 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461354 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461405 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461053 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:57 crc kubenswrapper[4872]: E1009 08:18:57.461831 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.461960 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:57 crc kubenswrapper[4872]: E1009 08:18:57.462959 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.564264 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.564315 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.564329 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.564347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.564359 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.667468 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.667544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.667556 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.667575 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.667589 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.770029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.770059 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.770072 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.770089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.770101 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.874986 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.875025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.875034 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.875048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.875057 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.977557 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.977623 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.977667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.977690 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:57 crc kubenswrapper[4872]: I1009 08:18:57.977706 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:57Z","lastTransitionTime":"2025-10-09T08:18:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.080163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.080287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.080304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.080321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.080333 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.183109 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.183189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.183206 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.183230 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.183297 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.285956 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.286000 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.286009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.286025 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.286033 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.388267 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.388304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.388312 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.388326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.388334 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.460933 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:18:58 crc kubenswrapper[4872]: E1009 08:18:58.461116 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.490359 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.490394 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.490402 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.490417 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.490425 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.593413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.593470 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.593483 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.593504 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.593517 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.696262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.696304 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.696316 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.696333 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.696345 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.799342 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.799389 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.799401 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.799420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.799433 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.903183 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.903245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.903265 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.903289 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:58 crc kubenswrapper[4872]: I1009 08:18:58.903308 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:58Z","lastTransitionTime":"2025-10-09T08:18:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.006538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.006592 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.006609 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.006656 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.006862 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.109278 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.109341 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.109357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.109382 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.109399 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.213060 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.213210 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.213231 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.213255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.213273 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.316089 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.316140 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.316163 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.316193 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.316214 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.406850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.406913 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.406927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.406946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.406960 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.423612 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:59Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.428137 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.428176 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.428186 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.428199 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.428208 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.445682 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:59Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.449791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.450222 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.450409 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.450558 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.450933 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.461131 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.461197 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.461350 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.461371 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.463043 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.463293 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.470243 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:59Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.474752 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.474798 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.474816 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.474833 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.474846 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.486057 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:59Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.489178 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.489234 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.489245 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.489260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.489270 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.500042 4872 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-09T08:18:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"4a6bfc25-1aa1-4792-9216-2dff0eba4a5b\\\",\\\"systemUUID\\\":\\\"40bf4641-1076-4716-a2ae-ad6807813135\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-09T08:18:59Z is after 2025-08-24T17:21:41Z" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.500266 4872 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.502013 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.502063 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.502074 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.502090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.502100 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.605522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.605580 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.605589 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.605614 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.605633 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.677635 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.677868 4872 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:59 crc kubenswrapper[4872]: E1009 08:18:59.677961 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs podName:326d4497-ac5f-4a64-afde-14ded51665b8 nodeName:}" failed. No retries permitted until 2025-10-09 08:20:03.677936572 +0000 UTC m=+161.868465258 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs") pod "network-metrics-daemon-8grb6" (UID: "326d4497-ac5f-4a64-afde-14ded51665b8") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.708542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.708607 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.708632 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.708688 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.708710 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.811787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.811820 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.811828 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.811841 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.811850 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.915044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.915078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.915090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.915107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:18:59 crc kubenswrapper[4872]: I1009 08:18:59.915119 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:18:59Z","lastTransitionTime":"2025-10-09T08:18:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.017522 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.017586 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.017604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.017627 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.017675 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.120260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.120331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.120355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.120387 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.120410 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.223536 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.223573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.223585 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.223602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.223614 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.326215 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.326271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.326283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.326302 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.326316 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.428815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.428854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.428864 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.428883 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.428902 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.461310 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:00 crc kubenswrapper[4872]: E1009 08:19:00.461474 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.530842 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.530872 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.530881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.530896 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.530905 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.634485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.634547 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.634571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.634622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.634677 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.738167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.738223 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.738237 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.738257 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.738269 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.841707 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.841760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.841792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.841830 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.841854 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.944262 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.944301 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.944311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.944326 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:00 crc kubenswrapper[4872]: I1009 08:19:00.944337 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:00Z","lastTransitionTime":"2025-10-09T08:19:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.046219 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.046265 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.046282 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.046311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.046325 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.149792 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.149858 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.149880 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.149902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.149923 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.252716 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.252793 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.252812 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.252838 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.252857 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.356603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.356709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.356727 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.356750 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.356766 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.458892 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.458946 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.458963 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.458988 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.459007 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.461038 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.461092 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.461123 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:01 crc kubenswrapper[4872]: E1009 08:19:01.461390 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:01 crc kubenswrapper[4872]: E1009 08:19:01.461505 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:01 crc kubenswrapper[4872]: E1009 08:19:01.461711 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.561787 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.561823 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.561865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.561881 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.561892 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.665347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.665424 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.665449 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.665692 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.665738 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.768379 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.768445 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.768462 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.768485 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.768501 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.871865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.871915 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.871928 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.871947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.871960 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.976048 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.976116 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.976142 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.976170 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:01 crc kubenswrapper[4872]: I1009 08:19:01.976191 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:01Z","lastTransitionTime":"2025-10-09T08:19:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.078850 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.078911 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.078930 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.078954 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.078972 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.181667 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.181733 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.181746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.181765 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.181778 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.284714 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.284771 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.284791 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.284813 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.284833 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.387744 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.387854 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.387891 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.387925 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.387947 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.460976 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:02 crc kubenswrapper[4872]: E1009 08:19:02.461199 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.462680 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:19:02 crc kubenswrapper[4872]: E1009 08:19:02.463075 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.491190 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.491321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.491347 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.491381 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.491406 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.496407 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=51.496348301 podStartE2EDuration="51.496348301s" podCreationTimestamp="2025-10-09 08:18:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.496288559 +0000 UTC m=+100.686817225" watchObservedRunningTime="2025-10-09 08:19:02.496348301 +0000 UTC m=+100.686876977" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.562135 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-7f65m" podStartSLOduration=81.56211106 podStartE2EDuration="1m21.56211106s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.529934462 +0000 UTC m=+100.720463158" watchObservedRunningTime="2025-10-09 08:19:02.56211106 +0000 UTC m=+100.752639726" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.562394 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-ssrfz" podStartSLOduration=81.562380329 podStartE2EDuration="1m21.562380329s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.560497171 +0000 UTC m=+100.751025817" watchObservedRunningTime="2025-10-09 08:19:02.562380329 +0000 UTC m=+100.752909025" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.593902 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.593976 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.594004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.594033 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.594056 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.644148 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=81.644129814 podStartE2EDuration="1m21.644129814s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.629563651 +0000 UTC m=+100.820092287" watchObservedRunningTime="2025-10-09 08:19:02.644129814 +0000 UTC m=+100.834658440" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.656373 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podStartSLOduration=81.656355555 podStartE2EDuration="1m21.656355555s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.655651044 +0000 UTC m=+100.846179680" watchObservedRunningTime="2025-10-09 08:19:02.656355555 +0000 UTC m=+100.846884181" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.696319 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.696361 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.696373 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.696388 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.696399 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.731742 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-bd2z9" podStartSLOduration=81.731721877 podStartE2EDuration="1m21.731721877s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.719214016 +0000 UTC m=+100.909742652" watchObservedRunningTime="2025-10-09 08:19:02.731721877 +0000 UTC m=+100.922250503" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.732402 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5ftzb" podStartSLOduration=81.73239851700001 podStartE2EDuration="1m21.732398517s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.731463789 +0000 UTC m=+100.921992415" watchObservedRunningTime="2025-10-09 08:19:02.732398517 +0000 UTC m=+100.922927143" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.743917 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=75.743899637 podStartE2EDuration="1m15.743899637s" podCreationTimestamp="2025-10-09 08:17:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.743491904 +0000 UTC m=+100.934020540" watchObservedRunningTime="2025-10-09 08:19:02.743899637 +0000 UTC m=+100.934428273" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.754898 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.754881721 podStartE2EDuration="16.754881721s" podCreationTimestamp="2025-10-09 08:18:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.75353501 +0000 UTC m=+100.944063636" watchObservedRunningTime="2025-10-09 08:19:02.754881721 +0000 UTC m=+100.945410357" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.777290 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-s7ffs" podStartSLOduration=80.777274381 podStartE2EDuration="1m20.777274381s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:02.776980952 +0000 UTC m=+100.967509578" watchObservedRunningTime="2025-10-09 08:19:02.777274381 +0000 UTC m=+100.967803007" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.798321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.798355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.798365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.798378 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.798387 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.900878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.900916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.900926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.900942 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:02 crc kubenswrapper[4872]: I1009 08:19:02.900953 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:02Z","lastTransitionTime":"2025-10-09T08:19:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.003004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.003287 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.003411 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.003516 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.003667 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.106189 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.106583 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.106596 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.106660 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.106676 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.208514 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.208564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.208574 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.208587 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.208596 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.311472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.311546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.311571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.311606 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.311627 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.414350 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.414399 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.414420 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.414442 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.414461 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.461433 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.461444 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:03 crc kubenswrapper[4872]: E1009 08:19:03.461680 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.461464 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:03 crc kubenswrapper[4872]: E1009 08:19:03.461883 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:03 crc kubenswrapper[4872]: E1009 08:19:03.461973 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.518357 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.518663 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.518754 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.518865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.518955 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.621192 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.621248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.621260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.621277 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.621289 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.723761 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.723818 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.723837 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.723857 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.723872 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.826395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.826450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.826471 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.826495 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.826511 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.928730 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.929021 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.929125 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.929224 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:03 crc kubenswrapper[4872]: I1009 08:19:03.929305 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:03Z","lastTransitionTime":"2025-10-09T08:19:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.032028 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.032071 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.032083 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.032098 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.032109 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.133926 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.133959 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.133967 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.133981 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.133992 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.236303 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.236331 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.236340 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.236353 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.236363 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.338776 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.339051 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.339120 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.339182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.339247 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.442397 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.442503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.442528 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.442561 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.442595 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.460899 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:04 crc kubenswrapper[4872]: E1009 08:19:04.461325 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.545413 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.545457 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.545474 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.545497 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.545520 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.648542 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.648594 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.648608 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.648629 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.648674 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.751539 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.751622 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.751674 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.751706 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.751728 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.854255 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.854346 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.854369 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.854395 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.854415 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.958090 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.958533 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.958747 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.958982 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:04 crc kubenswrapper[4872]: I1009 08:19:04.959184 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:04Z","lastTransitionTime":"2025-10-09T08:19:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.061540 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.061570 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.061577 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.061592 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.061600 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.164772 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.164815 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.164826 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.164843 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.164854 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.268746 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.268804 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.268817 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.268843 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.268855 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.371311 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.371355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.371365 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.371384 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.371396 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.460884 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.460919 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:05 crc kubenswrapper[4872]: E1009 08:19:05.461020 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.461031 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:05 crc kubenswrapper[4872]: E1009 08:19:05.461117 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:05 crc kubenswrapper[4872]: E1009 08:19:05.461197 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.473947 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.474029 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.474040 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.474056 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.474067 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.577431 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.577510 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.577529 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.577550 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.577569 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.680355 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.680441 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.680482 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.680513 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.680536 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.783574 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.783677 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.783703 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.783731 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.783753 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.887044 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.887366 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.887685 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.887906 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.888103 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.990801 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.991093 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.991172 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.991271 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:05 crc kubenswrapper[4872]: I1009 08:19:05.991350 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:05Z","lastTransitionTime":"2025-10-09T08:19:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.094796 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.094849 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.094865 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.094886 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.094900 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.198157 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.198218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.198232 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.198249 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.198263 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.300893 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.300957 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.300973 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.300994 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.301007 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.403386 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.403760 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.403927 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.404065 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.404251 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.461924 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:06 crc kubenswrapper[4872]: E1009 08:19:06.462140 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.481451 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.507545 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.508011 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.508248 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.508463 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.508685 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.611438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.611503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.611515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.611534 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.611549 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.714204 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.714503 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.714604 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.714709 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.714805 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.817705 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.817757 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.817774 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.817794 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.817810 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.921038 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.921079 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.921088 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.921104 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:06 crc kubenswrapper[4872]: I1009 08:19:06.921115 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:06Z","lastTransitionTime":"2025-10-09T08:19:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.024548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.024591 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.024602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.024617 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.024657 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.127564 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.127615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.127628 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.127675 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.127689 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.230367 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.230410 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.230419 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.230438 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.230447 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.333078 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.333147 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.333167 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.333196 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.333258 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.436775 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.436872 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.436892 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.436916 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.436931 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.461450 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.461492 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.461493 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:07 crc kubenswrapper[4872]: E1009 08:19:07.461660 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:07 crc kubenswrapper[4872]: E1009 08:19:07.461850 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:07 crc kubenswrapper[4872]: E1009 08:19:07.461942 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.539486 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.539546 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.539557 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.539579 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.539593 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.643415 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.643464 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.643475 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.643494 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.643527 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.747218 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.747276 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.747296 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.747314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.747327 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.851184 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.851260 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.851283 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.851314 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.851336 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.954847 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.954905 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.954924 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.954949 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:07 crc kubenswrapper[4872]: I1009 08:19:07.954967 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:07Z","lastTransitionTime":"2025-10-09T08:19:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.057741 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.058155 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.058372 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.058569 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.058807 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.162107 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.162472 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.162766 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.162941 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.163037 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.269440 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.269524 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.269538 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.269562 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.269578 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.372932 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.372984 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.372995 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.373017 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.373029 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.461845 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:08 crc kubenswrapper[4872]: E1009 08:19:08.462066 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.475450 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.475492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.475502 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.475515 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.475526 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.577811 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.577852 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.577861 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.577878 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.577889 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.680009 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.680061 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.680075 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.680095 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.680108 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.783392 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.783486 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.783507 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.783532 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.783552 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.887124 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.887172 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.887182 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.887197 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.887209 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.990004 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.990313 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.990492 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.990678 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:08 crc kubenswrapper[4872]: I1009 08:19:08.990830 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:08Z","lastTransitionTime":"2025-10-09T08:19:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.093469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.093549 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.093573 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.093597 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.093614 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.195939 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.195987 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.196001 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.196018 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.196030 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.299169 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.299277 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.299297 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.299321 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.299338 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.403085 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.403150 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.403172 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.403221 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.403244 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.461693 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.461771 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.461718 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:09 crc kubenswrapper[4872]: E1009 08:19:09.461921 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:09 crc kubenswrapper[4872]: E1009 08:19:09.462028 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:09 crc kubenswrapper[4872]: E1009 08:19:09.462189 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.505400 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.505447 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.505460 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.505480 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.505492 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.608544 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.608590 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.608602 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.608625 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.608658 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.711615 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.711698 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.711711 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.711742 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.711760 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.814469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.814548 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.814571 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.814603 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.814624 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.858390 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.858439 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.858452 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.858469 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.858483 4872 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-09T08:19:09Z","lastTransitionTime":"2025-10-09T08:19:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.908126 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf"] Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.908535 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.910780 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.911237 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.911409 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.911528 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.995632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aaaa47-69d4-4012-bf22-8d62961e12b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.996055 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12aaaa47-69d4-4012-bf22-8d62961e12b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.996247 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.996473 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:09 crc kubenswrapper[4872]: I1009 08:19:09.996608 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12aaaa47-69d4-4012-bf22-8d62961e12b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.097976 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aaaa47-69d4-4012-bf22-8d62961e12b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.098883 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12aaaa47-69d4-4012-bf22-8d62961e12b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.098945 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.099056 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.099115 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12aaaa47-69d4-4012-bf22-8d62961e12b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.099167 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.099263 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/12aaaa47-69d4-4012-bf22-8d62961e12b3-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.100562 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/12aaaa47-69d4-4012-bf22-8d62961e12b3-service-ca\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.109217 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12aaaa47-69d4-4012-bf22-8d62961e12b3-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.127712 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/12aaaa47-69d4-4012-bf22-8d62961e12b3-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-x2ctf\" (UID: \"12aaaa47-69d4-4012-bf22-8d62961e12b3\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.222577 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" Oct 09 08:19:10 crc kubenswrapper[4872]: I1009 08:19:10.460588 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:10 crc kubenswrapper[4872]: E1009 08:19:10.460924 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.060939 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" event={"ID":"12aaaa47-69d4-4012-bf22-8d62961e12b3","Type":"ContainerStarted","Data":"c482ec045051c6a6ce07fc8a948676b020ceab8a6e81c931a0d3cd8da7f0a427"} Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.061689 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" event={"ID":"12aaaa47-69d4-4012-bf22-8d62961e12b3","Type":"ContainerStarted","Data":"b6db108b5e2ba5b52d8b71252c0e5b612e827195cca5128ca2eeaef31cabc1b5"} Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.076088 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.076066724 podStartE2EDuration="5.076066724s" podCreationTimestamp="2025-10-09 08:19:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:09.94631448 +0000 UTC m=+108.136843116" watchObservedRunningTime="2025-10-09 08:19:11.076066724 +0000 UTC m=+109.266595350" Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.076707 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-x2ctf" podStartSLOduration=90.076698073 podStartE2EDuration="1m30.076698073s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:11.076223709 +0000 UTC m=+109.266752335" watchObservedRunningTime="2025-10-09 08:19:11.076698073 +0000 UTC m=+109.267226709" Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.460989 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:11 crc kubenswrapper[4872]: E1009 08:19:11.461352 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.461043 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:11 crc kubenswrapper[4872]: I1009 08:19:11.461000 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:11 crc kubenswrapper[4872]: E1009 08:19:11.461784 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:11 crc kubenswrapper[4872]: E1009 08:19:11.461659 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:12 crc kubenswrapper[4872]: I1009 08:19:12.461220 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:12 crc kubenswrapper[4872]: E1009 08:19:12.462510 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:13 crc kubenswrapper[4872]: I1009 08:19:13.460822 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:13 crc kubenswrapper[4872]: I1009 08:19:13.460842 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:13 crc kubenswrapper[4872]: I1009 08:19:13.460842 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:13 crc kubenswrapper[4872]: E1009 08:19:13.461302 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:13 crc kubenswrapper[4872]: E1009 08:19:13.461365 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:13 crc kubenswrapper[4872]: E1009 08:19:13.461173 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:14 crc kubenswrapper[4872]: I1009 08:19:14.461165 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:14 crc kubenswrapper[4872]: E1009 08:19:14.461505 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:15 crc kubenswrapper[4872]: I1009 08:19:15.461477 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:15 crc kubenswrapper[4872]: E1009 08:19:15.461793 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:15 crc kubenswrapper[4872]: I1009 08:19:15.461936 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:15 crc kubenswrapper[4872]: I1009 08:19:15.461977 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:15 crc kubenswrapper[4872]: E1009 08:19:15.462072 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:15 crc kubenswrapper[4872]: E1009 08:19:15.462258 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.084225 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/1.log" Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.085480 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/0.log" Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.085572 4872 generic.go:334] "Generic (PLEG): container finished" podID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" containerID="7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc" exitCode=1 Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.085632 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerDied","Data":"7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc"} Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.085715 4872 scope.go:117] "RemoveContainer" containerID="956b537012a451b3b04ad46d9a2ad6acf25322239e3a63fdafe4b18ad0b00b90" Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.086329 4872 scope.go:117] "RemoveContainer" containerID="7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc" Oct 09 08:19:16 crc kubenswrapper[4872]: E1009 08:19:16.086699 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5ftzb_openshift-multus(d5f349b9-7850-4b1b-86da-e89f7cfd32df)\"" pod="openshift-multus/multus-5ftzb" podUID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" Oct 09 08:19:16 crc kubenswrapper[4872]: I1009 08:19:16.461166 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:16 crc kubenswrapper[4872]: E1009 08:19:16.461341 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:17 crc kubenswrapper[4872]: I1009 08:19:17.090293 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/1.log" Oct 09 08:19:17 crc kubenswrapper[4872]: I1009 08:19:17.461009 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:17 crc kubenswrapper[4872]: I1009 08:19:17.461055 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:17 crc kubenswrapper[4872]: I1009 08:19:17.461067 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:17 crc kubenswrapper[4872]: E1009 08:19:17.461139 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:17 crc kubenswrapper[4872]: E1009 08:19:17.461225 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:17 crc kubenswrapper[4872]: E1009 08:19:17.461699 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:17 crc kubenswrapper[4872]: I1009 08:19:17.462147 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:19:17 crc kubenswrapper[4872]: E1009 08:19:17.462376 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-xnhdk_openshift-ovn-kubernetes(fac60b03-54bb-43e8-8994-3674d42b209b)\"" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" Oct 09 08:19:18 crc kubenswrapper[4872]: I1009 08:19:18.461332 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:18 crc kubenswrapper[4872]: E1009 08:19:18.462030 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:19 crc kubenswrapper[4872]: I1009 08:19:19.461570 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:19 crc kubenswrapper[4872]: I1009 08:19:19.461609 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:19 crc kubenswrapper[4872]: E1009 08:19:19.461801 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:19 crc kubenswrapper[4872]: I1009 08:19:19.461611 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:19 crc kubenswrapper[4872]: E1009 08:19:19.462013 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:19 crc kubenswrapper[4872]: E1009 08:19:19.462205 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:20 crc kubenswrapper[4872]: I1009 08:19:20.461086 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:20 crc kubenswrapper[4872]: E1009 08:19:20.461327 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:21 crc kubenswrapper[4872]: I1009 08:19:21.461161 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:21 crc kubenswrapper[4872]: E1009 08:19:21.461847 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:21 crc kubenswrapper[4872]: I1009 08:19:21.461168 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:21 crc kubenswrapper[4872]: E1009 08:19:21.462105 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:21 crc kubenswrapper[4872]: I1009 08:19:21.461161 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:21 crc kubenswrapper[4872]: E1009 08:19:21.462477 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:22 crc kubenswrapper[4872]: I1009 08:19:22.461493 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:22 crc kubenswrapper[4872]: E1009 08:19:22.465847 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:22 crc kubenswrapper[4872]: E1009 08:19:22.492578 4872 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 09 08:19:22 crc kubenswrapper[4872]: E1009 08:19:22.558873 4872 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 08:19:23 crc kubenswrapper[4872]: I1009 08:19:23.461592 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:23 crc kubenswrapper[4872]: E1009 08:19:23.461851 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:23 crc kubenswrapper[4872]: I1009 08:19:23.461942 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:23 crc kubenswrapper[4872]: I1009 08:19:23.462041 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:23 crc kubenswrapper[4872]: E1009 08:19:23.462168 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:23 crc kubenswrapper[4872]: E1009 08:19:23.462352 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:24 crc kubenswrapper[4872]: I1009 08:19:24.461519 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:24 crc kubenswrapper[4872]: E1009 08:19:24.461816 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:25 crc kubenswrapper[4872]: I1009 08:19:25.460686 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:25 crc kubenswrapper[4872]: I1009 08:19:25.460632 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:25 crc kubenswrapper[4872]: E1009 08:19:25.460909 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:25 crc kubenswrapper[4872]: I1009 08:19:25.460729 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:25 crc kubenswrapper[4872]: E1009 08:19:25.461033 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:25 crc kubenswrapper[4872]: E1009 08:19:25.461167 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:26 crc kubenswrapper[4872]: I1009 08:19:26.461527 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:26 crc kubenswrapper[4872]: E1009 08:19:26.461719 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:27 crc kubenswrapper[4872]: I1009 08:19:27.461498 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:27 crc kubenswrapper[4872]: I1009 08:19:27.461583 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:27 crc kubenswrapper[4872]: E1009 08:19:27.461627 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:27 crc kubenswrapper[4872]: I1009 08:19:27.461757 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:27 crc kubenswrapper[4872]: E1009 08:19:27.461776 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:27 crc kubenswrapper[4872]: E1009 08:19:27.461924 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:27 crc kubenswrapper[4872]: E1009 08:19:27.559699 4872 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 08:19:28 crc kubenswrapper[4872]: I1009 08:19:28.460604 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:28 crc kubenswrapper[4872]: E1009 08:19:28.460879 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:29 crc kubenswrapper[4872]: I1009 08:19:29.461163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:29 crc kubenswrapper[4872]: E1009 08:19:29.461313 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:29 crc kubenswrapper[4872]: I1009 08:19:29.461169 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:29 crc kubenswrapper[4872]: I1009 08:19:29.461163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:29 crc kubenswrapper[4872]: E1009 08:19:29.461706 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:29 crc kubenswrapper[4872]: E1009 08:19:29.461858 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:29 crc kubenswrapper[4872]: I1009 08:19:29.462113 4872 scope.go:117] "RemoveContainer" containerID="7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc" Oct 09 08:19:29 crc kubenswrapper[4872]: I1009 08:19:29.464426 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.136800 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/1.log" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.137234 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerStarted","Data":"6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd"} Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.139846 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/3.log" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.143097 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerStarted","Data":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.143701 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.283319 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podStartSLOduration=109.283296491 podStartE2EDuration="1m49.283296491s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:30.185440836 +0000 UTC m=+128.375969462" watchObservedRunningTime="2025-10-09 08:19:30.283296491 +0000 UTC m=+128.473825127" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.283655 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8grb6"] Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.283742 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:30 crc kubenswrapper[4872]: E1009 08:19:30.283835 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:30 crc kubenswrapper[4872]: I1009 08:19:30.461714 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:30 crc kubenswrapper[4872]: E1009 08:19:30.461968 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:31 crc kubenswrapper[4872]: I1009 08:19:31.460589 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:31 crc kubenswrapper[4872]: I1009 08:19:31.460619 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:31 crc kubenswrapper[4872]: I1009 08:19:31.460727 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:31 crc kubenswrapper[4872]: E1009 08:19:31.460874 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-8grb6" podUID="326d4497-ac5f-4a64-afde-14ded51665b8" Oct 09 08:19:31 crc kubenswrapper[4872]: E1009 08:19:31.461166 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 09 08:19:31 crc kubenswrapper[4872]: E1009 08:19:31.461326 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 09 08:19:32 crc kubenswrapper[4872]: I1009 08:19:32.460920 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:32 crc kubenswrapper[4872]: E1009 08:19:32.463074 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.461535 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.461681 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.462488 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.463879 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.464077 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.464230 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.464362 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.464485 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 09 08:19:33 crc kubenswrapper[4872]: I1009 08:19:33.465097 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 09 08:19:34 crc kubenswrapper[4872]: I1009 08:19:34.461243 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.432525 4872 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.482221 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.482748 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.483635 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rlfqc"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.484478 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.484763 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wcb54"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.485154 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.486511 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.486913 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.490403 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.490468 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.496739 4872 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.496801 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497000 4872 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497018 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497020 4872 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497084 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497152 4872 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.497321 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497589 4872 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497611 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497686 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497709 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.497758 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497775 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.498015 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.501331 4872 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.501383 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.501819 4872 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.501866 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.502355 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.497168 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.519615 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.519918 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.519956 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520246 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520280 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520333 4872 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520349 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520399 4872 reflector.go:561] object-"openshift-apiserver"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520414 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520457 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520469 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520545 4872 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520561 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520584 4872 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520601 4872 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520615 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520621 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520697 4872 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520739 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.520779 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520810 4872 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520824 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520864 4872 reflector.go:561] object-"openshift-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520782 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520877 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520884 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520916 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520929 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.520973 4872 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.520989 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.521022 4872 reflector.go:561] object-"openshift-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.521036 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: W1009 08:19:40.521234 4872 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Oct 09 08:19:40 crc kubenswrapper[4872]: E1009 08:19:40.521248 4872 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.521310 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.521938 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.522824 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.523116 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.523147 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.523872 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.524341 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.525219 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.526787 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.527209 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.527320 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.527410 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.528272 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.528408 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.528448 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.530939 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.531273 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.531600 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.532006 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.532536 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.532890 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.533036 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.533370 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.540731 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.541372 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bvcfj"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.541803 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.542748 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.553969 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.554746 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.555050 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.555322 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.555566 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.556936 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.557549 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dk2r4"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.558017 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.558890 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.559251 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.559628 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.561257 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.566085 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.566354 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.566492 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.566609 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.566880 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.567437 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.567816 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.567837 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.567976 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568258 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568367 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568458 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568581 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568755 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.568854 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569198 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569377 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569528 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569654 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569797 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569801 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.569925 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gsqtq"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.570362 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.570958 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.571508 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.571618 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.571891 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.573869 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.575284 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.587048 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.595203 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.595391 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.595837 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.598891 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.598966 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.604897 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606619 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606767 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606773 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606863 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606905 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606937 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606944 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607006 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607067 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607084 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.606918 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607169 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607567 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607622 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607691 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607806 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.607815 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.608029 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.608054 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.608297 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.608385 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.608895 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.609450 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.610542 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.610912 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.611242 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.612911 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.615218 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mqcvd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.615660 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.616276 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.616463 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.616767 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.616871 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.618131 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.619271 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.619863 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.625137 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.633707 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.634674 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.634914 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.635474 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.636938 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.637671 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.638215 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.639901 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.640292 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dt452"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.640656 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bkhr6"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.641056 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.641232 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.641373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.641980 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-86pkk"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.642295 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.642732 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fgzt2"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.643125 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.644412 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.644786 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.644935 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9rhp\" (UniqueName: \"kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645071 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44e52fef-7688-437b-8c44-af285a46273d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645344 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645510 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645631 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.645928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.646078 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpzch\" (UniqueName: \"kubernetes.io/projected/fd040384-e861-44c9-bffc-1dc10acb8e89-kube-api-access-tpzch\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.646258 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6slb\" (UniqueName: \"kubernetes.io/projected/8a5e4235-b5ba-495b-8462-8796c968d1f4-kube-api-access-w6slb\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.649456 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.656538 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.657689 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.660557 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.661802 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-nm6xv"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.662027 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.662364 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsngr\" (UniqueName: \"kubernetes.io/projected/44e52fef-7688-437b-8c44-af285a46273d-kube-api-access-xsngr\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.663437 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.664947 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.665001 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-audit-dir\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.665427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25be6ed5-a067-4cfa-bb26-98be70a39f54-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.665906 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.666904 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff0bb76-a914-4c38-90d2-f879b8c00b6e-serving-cert\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.667900 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668008 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.666982 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668706 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668726 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668755 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668786 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a5e4235-b5ba-495b-8462-8796c968d1f4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668812 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668835 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr8vr\" (UniqueName: \"kubernetes.io/projected/6f555e5b-289d-40e7-9671-c47c69282416-kube-api-access-qr8vr\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668856 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668880 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-node-pullsecrets\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668905 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668927 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-trusted-ca\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668952 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.668993 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.669019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcxk6\" (UniqueName: \"kubernetes.io/projected/cff0bb76-a914-4c38-90d2-f879b8c00b6e-kube-api-access-dcxk6\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.669299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25be6ed5-a067-4cfa-bb26-98be70a39f54-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.669342 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.669669 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-dir\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.669706 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5e4235-b5ba-495b-8462-8796c968d1f4-serving-cert\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.670010 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5r56c\" (UniqueName: \"kubernetes.io/projected/25be6ed5-a067-4cfa-bb26-98be70a39f54-kube-api-access-5r56c\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.670248 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.670491 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.670527 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dx2x\" (UniqueName: \"kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.670986 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671017 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671139 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671161 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671318 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671340 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-config\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.671570 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvcrp\" (UniqueName: \"kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.672836 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.673549 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.674297 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.677357 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.679298 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.683668 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rlfqc"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.683793 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.689925 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.701594 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.704225 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dk2r4"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.705795 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.707546 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.707648 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.710840 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bvcfj"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.713426 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.717231 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wcb54"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.755286 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.756493 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.756976 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.757740 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.760758 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bkhr6"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.761795 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.763161 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.763341 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.764655 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.764784 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.765701 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.766845 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.768879 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.770264 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mqcvd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.771400 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.772972 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.773694 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61adb101-0d3a-4018-aa4f-4ed4a435148b-service-ca-bundle\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774092 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774111 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774135 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774156 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdkn2\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-kube-api-access-fdkn2\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774174 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff0bb76-a914-4c38-90d2-f879b8c00b6e-serving-cert\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774189 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774206 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035a7dd6-e488-4af1-ae0f-345f6232751f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774224 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774239 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a5e4235-b5ba-495b-8462-8796c968d1f4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774256 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-apiservice-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774300 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-metrics-certs\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774315 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-client\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774331 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5q6p\" (UniqueName: \"kubernetes.io/projected/159cd083-1078-4cb3-9142-b985c7a53823-kube-api-access-v5q6p\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774350 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-node-pullsecrets\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774363 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr8vr\" (UniqueName: \"kubernetes.io/projected/6f555e5b-289d-40e7-9671-c47c69282416-kube-api-access-qr8vr\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774379 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774395 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bflb\" (UniqueName: \"kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774409 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bb6c618e-f78b-40eb-8556-3824be52910d-metrics-tls\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774424 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774439 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-trusted-ca\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774456 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774473 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774489 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774513 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/81d1c103-d6b5-4e05-a262-10a658db5948-metrics-tls\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774531 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-service-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774550 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25be6ed5-a067-4cfa-bb26-98be70a39f54-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774568 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75r57\" (UniqueName: \"kubernetes.io/projected/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-kube-api-access-75r57\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774585 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5r56c\" (UniqueName: \"kubernetes.io/projected/25be6ed5-a067-4cfa-bb26-98be70a39f54-kube-api-access-5r56c\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774885 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774918 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774936 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dx2x\" (UniqueName: \"kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774956 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035a7dd6-e488-4af1-ae0f-345f6232751f-config\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774980 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.774997 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775015 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775029 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-stats-auth\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775045 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775062 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775077 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11d63870-f531-4733-a134-6de125f2a7a5-machine-approver-tls\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grdkm\" (UniqueName: \"kubernetes.io/projected/36a1a915-0f81-4923-a700-98ca7584e702-kube-api-access-grdkm\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775108 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775130 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775150 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-serving-cert\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775166 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvcrp\" (UniqueName: \"kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775204 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7af426f-a554-434f-9735-922bebb651bf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775219 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9rhp\" (UniqueName: \"kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775252 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775268 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/847c2d27-b3a4-4dcb-a00e-307360b85f59-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775286 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775303 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-serving-cert\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775319 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775338 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775368 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpzch\" (UniqueName: \"kubernetes.io/projected/fd040384-e861-44c9-bffc-1dc10acb8e89-kube-api-access-tpzch\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775390 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-profile-collector-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775405 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775423 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6slb\" (UniqueName: \"kubernetes.io/projected/8a5e4235-b5ba-495b-8462-8796c968d1f4-kube-api-access-w6slb\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775437 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775452 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsngr\" (UniqueName: \"kubernetes.io/projected/44e52fef-7688-437b-8c44-af285a46273d-kube-api-access-xsngr\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775475 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dmv5\" (UniqueName: \"kubernetes.io/projected/61adb101-0d3a-4018-aa4f-4ed4a435148b-kube-api-access-8dmv5\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775509 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775527 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-audit-dir\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775542 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjkxb\" (UniqueName: \"kubernetes.io/projected/f9b56021-cd13-4c6a-9f73-41db5b77d85b-kube-api-access-hjkxb\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775557 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6lc\" (UniqueName: \"kubernetes.io/projected/1b2002e1-bd93-447f-9a8b-e647308be981-kube-api-access-9n6lc\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775573 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzfmh\" (UniqueName: \"kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775590 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25be6ed5-a067-4cfa-bb26-98be70a39f54-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbb79\" (UniqueName: \"kubernetes.io/projected/11d63870-f531-4733-a134-6de125f2a7a5-kube-api-access-cbb79\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775623 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775663 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lpwv\" (UniqueName: \"kubernetes.io/projected/982bf7fe-335b-4fb6-be38-62de01442760-kube-api-access-4lpwv\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775680 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-config\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775696 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7z9x4\" (UniqueName: \"kubernetes.io/projected/4924a504-d8b5-406f-8379-19db356f0ba0-kube-api-access-7z9x4\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775712 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775727 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775743 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775762 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775780 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775795 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-srv-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775809 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775824 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775838 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/03a6afc5-c556-4b61-a41a-c6004edc386c-tmpfs\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775852 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-proxy-tls\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775887 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775931 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.775896 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36a1a915-0f81-4923-a700-98ca7584e702-cert\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776212 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-auth-proxy-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776235 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slgwk\" (UniqueName: \"kubernetes.io/projected/4fb7cde4-b48a-4b5d-8352-54b49a017713-kube-api-access-slgwk\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776269 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcxk6\" (UniqueName: \"kubernetes.io/projected/cff0bb76-a914-4c38-90d2-f879b8c00b6e-kube-api-access-dcxk6\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776288 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776350 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776377 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-dir\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776400 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847c2d27-b3a4-4dcb-a00e-307360b85f59-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776420 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-auth-proxy-config\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776436 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmd4g\" (UniqueName: \"kubernetes.io/projected/1ec987d5-2185-4e9e-96bc-307ea83ad0d0-kube-api-access-rmd4g\") pod \"migrator-59844c95c7-l275r\" (UID: \"1ec987d5-2185-4e9e-96bc-307ea83ad0d0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776454 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5e4235-b5ba-495b-8462-8796c968d1f4-serving-cert\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776484 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb6c618e-f78b-40eb-8556-3824be52910d-trusted-ca\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776501 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccfcz\" (UniqueName: \"kubernetes.io/projected/03a6afc5-c556-4b61-a41a-c6004edc386c-kube-api-access-ccfcz\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776520 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-default-certificate\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776564 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776581 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776600 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/847c2d27-b3a4-4dcb-a00e-307360b85f59-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776615 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776630 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776662 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-config\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776685 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159cd083-1078-4cb3-9142-b985c7a53823-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159cd083-1078-4cb3-9142-b985c7a53823-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776720 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44e52fef-7688-437b-8c44-af285a46273d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776737 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776754 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sblhk\" (UniqueName: \"kubernetes.io/projected/e7af426f-a554-434f-9735-922bebb651bf-kube-api-access-sblhk\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776770 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776787 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hm2p2\" (UniqueName: \"kubernetes.io/projected/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-kube-api-access-hm2p2\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776803 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-webhook-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776822 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-config\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776839 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776864 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035a7dd6-e488-4af1-ae0f-345f6232751f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.776881 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7dsf\" (UniqueName: \"kubernetes.io/projected/81d1c103-d6b5-4e05-a262-10a658db5948-kube-api-access-b7dsf\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.779556 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8a5e4235-b5ba-495b-8462-8796c968d1f4-available-featuregates\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.779665 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-dir\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.781177 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cff0bb76-a914-4c38-90d2-f879b8c00b6e-serving-cert\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.781720 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-audit-dir\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.781776 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6f555e5b-289d-40e7-9671-c47c69282416-node-pullsecrets\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.782552 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25be6ed5-a067-4cfa-bb26-98be70a39f54-config\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.782652 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.782677 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.785618 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-trusted-ca\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.781624 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.786112 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25be6ed5-a067-4cfa-bb26-98be70a39f54-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.786254 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.786390 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.787250 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.787802 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cff0bb76-a914-4c38-90d2-f879b8c00b6e-config\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.788377 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/44e52fef-7688-437b-8c44-af285a46273d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.788757 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8a5e4235-b5ba-495b-8462-8796c968d1f4-serving-cert\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.789269 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.789326 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-86pkk"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.790144 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-9dqkl"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.791422 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.791535 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.792526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dt452"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.797767 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fgzt2"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.798956 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.801856 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.804831 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nm6xv"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.805671 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.805863 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.806947 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.807886 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.809023 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.809878 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.811046 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-dfq2v"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.812191 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.812945 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbd5f"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.813827 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.815675 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dfq2v"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.817125 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbd5f"] Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.825737 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.865588 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877625 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877672 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grdkm\" (UniqueName: \"kubernetes.io/projected/36a1a915-0f81-4923-a700-98ca7584e702-kube-api-access-grdkm\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877696 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-serving-cert\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877712 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877734 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7af426f-a554-434f-9735-922bebb651bf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877750 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877781 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/847c2d27-b3a4-4dcb-a00e-307360b85f59-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877793 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.877811 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-serving-cert\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878009 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-profile-collector-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878041 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878068 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878126 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dmv5\" (UniqueName: \"kubernetes.io/projected/61adb101-0d3a-4018-aa4f-4ed4a435148b-kube-api-access-8dmv5\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878155 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjkxb\" (UniqueName: \"kubernetes.io/projected/f9b56021-cd13-4c6a-9f73-41db5b77d85b-kube-api-access-hjkxb\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6lc\" (UniqueName: \"kubernetes.io/projected/1b2002e1-bd93-447f-9a8b-e647308be981-kube-api-access-9n6lc\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878209 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzfmh\" (UniqueName: \"kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878234 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbb79\" (UniqueName: \"kubernetes.io/projected/11d63870-f531-4733-a134-6de125f2a7a5-kube-api-access-cbb79\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878264 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7z9x4\" (UniqueName: \"kubernetes.io/projected/4924a504-d8b5-406f-8379-19db356f0ba0-kube-api-access-7z9x4\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878289 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878312 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878337 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lpwv\" (UniqueName: \"kubernetes.io/projected/982bf7fe-335b-4fb6-be38-62de01442760-kube-api-access-4lpwv\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878360 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-config\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878412 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878438 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878465 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878494 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-srv-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878517 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878543 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878567 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-proxy-tls\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878620 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/03a6afc5-c556-4b61-a41a-c6004edc386c-tmpfs\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878660 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36a1a915-0f81-4923-a700-98ca7584e702-cert\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878703 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-auth-proxy-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878730 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slgwk\" (UniqueName: \"kubernetes.io/projected/4fb7cde4-b48a-4b5d-8352-54b49a017713-kube-api-access-slgwk\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878771 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-auth-proxy-config\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878795 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmd4g\" (UniqueName: \"kubernetes.io/projected/1ec987d5-2185-4e9e-96bc-307ea83ad0d0-kube-api-access-rmd4g\") pod \"migrator-59844c95c7-l275r\" (UID: \"1ec987d5-2185-4e9e-96bc-307ea83ad0d0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878818 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878853 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847c2d27-b3a4-4dcb-a00e-307360b85f59-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878892 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccfcz\" (UniqueName: \"kubernetes.io/projected/03a6afc5-c556-4b61-a41a-c6004edc386c-kube-api-access-ccfcz\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878915 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb6c618e-f78b-40eb-8556-3824be52910d-trusted-ca\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878938 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-default-certificate\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.878980 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879015 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879033 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879062 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/847c2d27-b3a4-4dcb-a00e-307360b85f59-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879086 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879111 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159cd083-1078-4cb3-9142-b985c7a53823-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159cd083-1078-4cb3-9142-b985c7a53823-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879164 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879192 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sblhk\" (UniqueName: \"kubernetes.io/projected/e7af426f-a554-434f-9735-922bebb651bf-kube-api-access-sblhk\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879229 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hm2p2\" (UniqueName: \"kubernetes.io/projected/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-kube-api-access-hm2p2\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879251 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-webhook-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879274 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-config\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879299 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035a7dd6-e488-4af1-ae0f-345f6232751f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879541 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7dsf\" (UniqueName: \"kubernetes.io/projected/81d1c103-d6b5-4e05-a262-10a658db5948-kube-api-access-b7dsf\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879569 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61adb101-0d3a-4018-aa4f-4ed4a435148b-service-ca-bundle\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879689 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879708 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-auth-proxy-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.879715 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.880627 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb6c618e-f78b-40eb-8556-3824be52910d-trusted-ca\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881191 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/03a6afc5-c556-4b61-a41a-c6004edc386c-tmpfs\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881391 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881425 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-auth-proxy-config\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881936 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/11d63870-f531-4733-a134-6de125f2a7a5-config\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.881997 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882051 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdkn2\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-kube-api-access-fdkn2\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882162 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035a7dd6-e488-4af1-ae0f-345f6232751f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882192 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-client\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882211 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5q6p\" (UniqueName: \"kubernetes.io/projected/159cd083-1078-4cb3-9142-b985c7a53823-kube-api-access-v5q6p\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882214 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/159cd083-1078-4cb3-9142-b985c7a53823-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882262 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-apiservice-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/61adb101-0d3a-4018-aa4f-4ed4a435148b-service-ca-bundle\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882299 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-metrics-certs\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882318 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bflb\" (UniqueName: \"kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882709 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bb6c618e-f78b-40eb-8556-3824be52910d-metrics-tls\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882808 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882813 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/81d1c103-d6b5-4e05-a262-10a658db5948-metrics-tls\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882852 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-service-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882879 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75r57\" (UniqueName: \"kubernetes.io/projected/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-kube-api-access-75r57\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.882909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035a7dd6-e488-4af1-ae0f-345f6232751f-config\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883069 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883099 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883125 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-stats-auth\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883166 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883188 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11d63870-f531-4733-a134-6de125f2a7a5-machine-approver-tls\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883758 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.883163 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.884050 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/035a7dd6-e488-4af1-ae0f-345f6232751f-config\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.884453 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.885062 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.885437 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.885684 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-default-certificate\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.886007 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/11d63870-f531-4733-a134-6de125f2a7a5-machine-approver-tls\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.886539 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-stats-auth\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.886762 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.886875 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/61adb101-0d3a-4018-aa4f-4ed4a435148b-metrics-certs\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.886993 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.887027 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/159cd083-1078-4cb3-9142-b985c7a53823-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.887699 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/bb6c618e-f78b-40eb-8556-3824be52910d-metrics-tls\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.890464 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.890853 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.901652 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.905470 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.916817 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/035a7dd6-e488-4af1-ae0f-345f6232751f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.925304 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.944989 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.952198 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-proxy-tls\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.964687 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 09 08:19:40 crc kubenswrapper[4872]: I1009 08:19:40.984685 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.004661 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.024610 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.031535 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-serving-cert\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.044681 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.064588 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.079925 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-config\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.085393 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.094880 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-service-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.104982 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.114211 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-serving-cert\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.125054 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.136735 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-client\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.144131 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.164996 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.184891 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.204816 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.212720 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-config\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.224560 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.233078 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/1b2002e1-bd93-447f-9a8b-e647308be981-etcd-ca\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.245589 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.264831 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.284559 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.296886 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-apiservice-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.297013 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/03a6afc5-c556-4b61-a41a-c6004edc386c-webhook-cert\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.306090 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.325633 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.333615 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-profile-collector-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.338711 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.345264 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.354709 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/982bf7fe-335b-4fb6-be38-62de01442760-srv-cert\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.365841 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.385731 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.404888 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.425164 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.444387 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.471935 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.485594 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.505579 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.524864 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.546429 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.552209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e7af426f-a554-434f-9735-922bebb651bf-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.565380 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.576325 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/81d1c103-d6b5-4e05-a262-10a658db5948-metrics-tls\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.585126 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.605852 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.624520 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.643029 4872 request.go:700] Waited for 1.001434758s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/secrets?fieldSelector=metadata.name%3Dkube-scheduler-operator-serving-cert&limit=500&resourceVersion=0 Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.644984 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.655324 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/847c2d27-b3a4-4dcb-a00e-307360b85f59-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.664596 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.671845 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/847c2d27-b3a4-4dcb-a00e-307360b85f59-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.684602 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.704206 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.724826 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.744388 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.757404 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/36a1a915-0f81-4923-a700-98ca7584e702-cert\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.765802 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.775531 4872 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.775611 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.275587473 +0000 UTC m=+140.466116109 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.776836 4872 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.776893 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.276879132 +0000 UTC m=+140.467407778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.777086 4872 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.777377 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.277290815 +0000 UTC m=+140.467819481 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779299 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779337 4872 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779390 4872 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779408 4872 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779479 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779667 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.279621216 +0000 UTC m=+140.470149922 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779889 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.279874154 +0000 UTC m=+140.470402870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779921 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.279907005 +0000 UTC m=+140.470435741 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779946 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.279936276 +0000 UTC m=+140.470465012 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.779970 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.279961156 +0000 UTC m=+140.470489892 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780698 4872 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780706 4872 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780742 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.28073127 +0000 UTC m=+140.471259906 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780763 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780778 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780810 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.280801982 +0000 UTC m=+140.471330618 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780847 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.280819263 +0000 UTC m=+140.471347939 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.780888 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.280869134 +0000 UTC m=+140.471397870 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.781949 4872 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.782034 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.282009879 +0000 UTC m=+140.472538545 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.782842 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.782902 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.282885255 +0000 UTC m=+140.473413971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.782933 4872 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.782972 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.282961268 +0000 UTC m=+140.473490004 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.783120 4872 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.783131 4872 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.783168 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.283155044 +0000 UTC m=+140.473683750 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.783243 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.283220666 +0000 UTC m=+140.473749362 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.785604 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.805888 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.825063 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.844459 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.865010 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.877768 4872 secret.go:188] Couldn't get secret openshift-operator-lifecycle-manager/olm-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.877843 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert podName:f9b56021-cd13-4c6a-9f73-41db5b77d85b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.377823349 +0000 UTC m=+140.568351975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "srv-cert" (UniqueName: "kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert") pod "olm-operator-6b444d44fb-mqrzg" (UID: "f9b56021-cd13-4c6a-9f73-41db5b77d85b") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.879816 4872 secret.go:188] Couldn't get secret openshift-machine-config-operator/mco-proxy-tls: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.879879 4872 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/machine-config-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.879904 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls podName:4fb7cde4-b48a-4b5d-8352-54b49a017713 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.37985442 +0000 UTC m=+140.570383046 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls") pod "machine-config-operator-74547568cd-28xtq" (UID: "4fb7cde4-b48a-4b5d-8352-54b49a017713") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.879922 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images podName:4fb7cde4-b48a-4b5d-8352-54b49a017713 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.379914552 +0000 UTC m=+140.570443178 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images") pod "machine-config-operator-74547568cd-28xtq" (UID: "4fb7cde4-b48a-4b5d-8352-54b49a017713") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.882629 4872 secret.go:188] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.882685 4872 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.882695 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert podName:4924a504-d8b5-406f-8379-19db356f0ba0 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.382681397 +0000 UTC m=+140.573210023 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert") pod "kube-storage-version-migrator-operator-b67b599dd-mlfjj" (UID: "4924a504-d8b5-406f-8379-19db356f0ba0") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: E1009 08:19:41.882725 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config podName:4924a504-d8b5-406f-8379-19db356f0ba0 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:42.382715828 +0000 UTC m=+140.573244454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config") pod "kube-storage-version-migrator-operator-b67b599dd-mlfjj" (UID: "4924a504-d8b5-406f-8379-19db356f0ba0") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.884239 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.904732 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.926345 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.945313 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.965331 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 09 08:19:41 crc kubenswrapper[4872]: I1009 08:19:41.985568 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.005283 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.025673 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.045989 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.064979 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.085162 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.105399 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.125253 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.144876 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.165773 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.185965 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.206083 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.226072 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.252231 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.265576 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.287414 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.305285 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308059 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308231 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308374 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308434 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308490 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308547 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308579 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308721 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308794 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308855 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.308947 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.309052 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.309289 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.309371 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.309448 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.335073 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.344584 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.363967 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.410347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.410389 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.410497 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.410597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.410739 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.411443 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4fb7cde4-b48a-4b5d-8352-54b49a017713-images\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.411977 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4924a504-d8b5-406f-8379-19db356f0ba0-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.414013 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4fb7cde4-b48a-4b5d-8352-54b49a017713-proxy-tls\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.414116 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f9b56021-cd13-4c6a-9f73-41db5b77d85b-srv-cert\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.416050 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4924a504-d8b5-406f-8379-19db356f0ba0-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.422474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5r56c\" (UniqueName: \"kubernetes.io/projected/25be6ed5-a067-4cfa-bb26-98be70a39f54-kube-api-access-5r56c\") pod \"openshift-apiserver-operator-796bbdcf4f-5lh66\" (UID: \"25be6ed5-a067-4cfa-bb26-98be70a39f54\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.438790 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.446726 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dx2x\" (UniqueName: \"kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x\") pod \"controller-manager-879f6c89f-hfxc8\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.462041 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcxk6\" (UniqueName: \"kubernetes.io/projected/cff0bb76-a914-4c38-90d2-f879b8c00b6e-kube-api-access-dcxk6\") pod \"console-operator-58897d9998-bvcfj\" (UID: \"cff0bb76-a914-4c38-90d2-f879b8c00b6e\") " pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.522688 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9rhp\" (UniqueName: \"kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp\") pod \"route-controller-manager-6576b87f9c-pxn8j\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.547626 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6slb\" (UniqueName: \"kubernetes.io/projected/8a5e4235-b5ba-495b-8462-8796c968d1f4-kube-api-access-w6slb\") pod \"openshift-config-operator-7777fb866f-5xj5z\" (UID: \"8a5e4235-b5ba-495b-8462-8796c968d1f4\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.560286 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsngr\" (UniqueName: \"kubernetes.io/projected/44e52fef-7688-437b-8c44-af285a46273d-kube-api-access-xsngr\") pod \"cluster-samples-operator-665b6dd947-fv2jp\" (UID: \"44e52fef-7688-437b-8c44-af285a46273d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.584715 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.604361 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.623270 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66"] Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.624743 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.631244 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.643199 4872 request.go:700] Waited for 1.830741319s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.645018 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.665028 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.684981 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.705861 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.713929 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.725555 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.744732 4872 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.749911 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.762922 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.773106 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.785121 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.813162 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/847c2d27-b3a4-4dcb-a00e-307360b85f59-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-pn8h7\" (UID: \"847c2d27-b3a4-4dcb-a00e-307360b85f59\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.829298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grdkm\" (UniqueName: \"kubernetes.io/projected/36a1a915-0f81-4923-a700-98ca7584e702-kube-api-access-grdkm\") pod \"ingress-canary-86pkk\" (UID: \"36a1a915-0f81-4923-a700-98ca7584e702\") " pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:42 crc kubenswrapper[4872]: W1009 08:19:42.832825 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09a17519_bda3_4246_bff9_adf452b800b0.slice/crio-971e528376523222edae19f0a85b84c9297a639bb77730595f5e9ef79ebb8f27 WatchSource:0}: Error finding container 971e528376523222edae19f0a85b84c9297a639bb77730595f5e9ef79ebb8f27: Status 404 returned error can't find the container with id 971e528376523222edae19f0a85b84c9297a639bb77730595f5e9ef79ebb8f27 Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.844285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzfmh\" (UniqueName: \"kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh\") pod \"marketplace-operator-79b997595-xlstb\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.865941 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjkxb\" (UniqueName: \"kubernetes.io/projected/f9b56021-cd13-4c6a-9f73-41db5b77d85b-kube-api-access-hjkxb\") pod \"olm-operator-6b444d44fb-mqrzg\" (UID: \"f9b56021-cd13-4c6a-9f73-41db5b77d85b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.880135 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6lc\" (UniqueName: \"kubernetes.io/projected/1b2002e1-bd93-447f-9a8b-e647308be981-kube-api-access-9n6lc\") pod \"etcd-operator-b45778765-mqcvd\" (UID: \"1b2002e1-bd93-447f-9a8b-e647308be981\") " pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.937931 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbb79\" (UniqueName: \"kubernetes.io/projected/11d63870-f531-4733-a134-6de125f2a7a5-kube-api-access-cbb79\") pod \"machine-approver-56656f9798-cs9zz\" (UID: \"11d63870-f531-4733-a134-6de125f2a7a5\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.940208 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.940599 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lpwv\" (UniqueName: \"kubernetes.io/projected/982bf7fe-335b-4fb6-be38-62de01442760-kube-api-access-4lpwv\") pod \"catalog-operator-68c6474976-978jd\" (UID: \"982bf7fe-335b-4fb6-be38-62de01442760\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.949379 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7z9x4\" (UniqueName: \"kubernetes.io/projected/4924a504-d8b5-406f-8379-19db356f0ba0-kube-api-access-7z9x4\") pod \"kube-storage-version-migrator-operator-b67b599dd-mlfjj\" (UID: \"4924a504-d8b5-406f-8379-19db356f0ba0\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.955931 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.958331 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.962620 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmd4g\" (UniqueName: \"kubernetes.io/projected/1ec987d5-2185-4e9e-96bc-307ea83ad0d0-kube-api-access-rmd4g\") pod \"migrator-59844c95c7-l275r\" (UID: \"1ec987d5-2185-4e9e-96bc-307ea83ad0d0\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" Oct 09 08:19:42 crc kubenswrapper[4872]: W1009 08:19:42.975307 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod953119f6_0fbb_41b4_9901_dfe2dfdcbc98.slice/crio-954442ecb5d7023f091a899fcf7f0d40f18c5500ecd7774d610d9104373f10ab WatchSource:0}: Error finding container 954442ecb5d7023f091a899fcf7f0d40f18c5500ecd7774d610d9104373f10ab: Status 404 returned error can't find the container with id 954442ecb5d7023f091a899fcf7f0d40f18c5500ecd7774d610d9104373f10ab Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.978936 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slgwk\" (UniqueName: \"kubernetes.io/projected/4fb7cde4-b48a-4b5d-8352-54b49a017713-kube-api-access-slgwk\") pod \"machine-config-operator-74547568cd-28xtq\" (UID: \"4fb7cde4-b48a-4b5d-8352-54b49a017713\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:42 crc kubenswrapper[4872]: I1009 08:19:42.999415 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dmv5\" (UniqueName: \"kubernetes.io/projected/61adb101-0d3a-4018-aa4f-4ed4a435148b-kube-api-access-8dmv5\") pod \"router-default-5444994796-gsqtq\" (UID: \"61adb101-0d3a-4018-aa4f-4ed4a435148b\") " pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.016206 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.026919 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.032997 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sblhk\" (UniqueName: \"kubernetes.io/projected/e7af426f-a554-434f-9735-922bebb651bf-kube-api-access-sblhk\") pod \"multus-admission-controller-857f4d67dd-dt452\" (UID: \"e7af426f-a554-434f-9735-922bebb651bf\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.041805 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-86pkk" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.055201 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccfcz\" (UniqueName: \"kubernetes.io/projected/03a6afc5-c556-4b61-a41a-c6004edc386c-kube-api-access-ccfcz\") pod \"packageserver-d55dfcdfc-h2ll5\" (UID: \"03a6afc5-c556-4b61-a41a-c6004edc386c\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.056828 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.065977 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.069104 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hm2p2\" (UniqueName: \"kubernetes.io/projected/7b2f7d7a-26a5-4566-916c-cf5af4d0cec6-kube-api-access-hm2p2\") pod \"service-ca-operator-777779d784-2lf9z\" (UID: \"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.072601 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.082389 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/035a7dd6-e488-4af1-ae0f-345f6232751f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-54wdx\" (UID: \"035a7dd6-e488-4af1-ae0f-345f6232751f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.104427 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdkn2\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-kube-api-access-fdkn2\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.117104 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.128516 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bflb\" (UniqueName: \"kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb\") pod \"oauth-openshift-558db77b4-g747n\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.153531 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5q6p\" (UniqueName: \"kubernetes.io/projected/159cd083-1078-4cb3-9142-b985c7a53823-kube-api-access-v5q6p\") pod \"openshift-controller-manager-operator-756b6f6bc6-lm8n7\" (UID: \"159cd083-1078-4cb3-9142-b985c7a53823\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.161989 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.165334 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.171036 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7dsf\" (UniqueName: \"kubernetes.io/projected/81d1c103-d6b5-4e05-a262-10a658db5948-kube-api-access-b7dsf\") pod \"dns-operator-744455d44c-bkhr6\" (UID: \"81d1c103-d6b5-4e05-a262-10a658db5948\") " pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.171295 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.172061 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.177601 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.183212 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bb6c618e-f78b-40eb-8556-3824be52910d-bound-sa-token\") pod \"ingress-operator-5b745b69d9-g2q4g\" (UID: \"bb6c618e-f78b-40eb-8556-3824be52910d\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.201900 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75r57\" (UniqueName: \"kubernetes.io/projected/3be12f5f-7fdb-4d42-b66e-42bccc2980e7-kube-api-access-75r57\") pod \"machine-config-controller-84d6567774-5zq9q\" (UID: \"3be12f5f-7fdb-4d42-b66e-42bccc2980e7\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.206421 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.212470 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.220513 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.223014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" event={"ID":"8a5e4235-b5ba-495b-8462-8796c968d1f4","Type":"ContainerStarted","Data":"be344028f650d6fd7fa0dd16a6e1f066cb8c30ac49544fdaebcb27b3cd3f3526"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.226331 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.228984 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.231031 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.242248 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.242361 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" event={"ID":"25be6ed5-a067-4cfa-bb26-98be70a39f54","Type":"ContainerStarted","Data":"6406fb4deae649dfd37b8b0311a0ff25a88b44e2ba271ddba81c3eee8f562334"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.242420 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" event={"ID":"25be6ed5-a067-4cfa-bb26-98be70a39f54","Type":"ContainerStarted","Data":"98352d166599a2599c02b1c139b3526afe7a98066558e21569bdd7293c6a297d"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.248709 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.250713 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.260619 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" event={"ID":"953119f6-0fbb-41b4-9901-dfe2dfdcbc98","Type":"ContainerStarted","Data":"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.260700 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" event={"ID":"953119f6-0fbb-41b4-9901-dfe2dfdcbc98","Type":"ContainerStarted","Data":"954442ecb5d7023f091a899fcf7f0d40f18c5500ecd7774d610d9104373f10ab"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.261030 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.265771 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-mqcvd"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.268280 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-encryption-config\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.268410 4872 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-pxn8j container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" start-of-body= Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.268455 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.27:8443/healthz\": dial tcp 10.217.0.27:8443: connect: connection refused" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.269363 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-bvcfj"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.272711 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.272884 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.273409 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" event={"ID":"09a17519-bda3-4246-bff9-adf452b800b0","Type":"ContainerStarted","Data":"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.273457 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" event={"ID":"09a17519-bda3-4246-bff9-adf452b800b0","Type":"ContainerStarted","Data":"971e528376523222edae19f0a85b84c9297a639bb77730595f5e9ef79ebb8f27"} Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.274155 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.283085 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.285327 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.287550 4872 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-hfxc8 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.287594 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" podUID="09a17519-bda3-4246-bff9-adf452b800b0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.308763 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.308839 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.308816576 +0000 UTC m=+142.499345202 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.308958 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309135 4872 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309146 4872 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309195 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309172327 +0000 UTC m=+142.499701063 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309205 4872 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309218 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309208508 +0000 UTC m=+142.499737264 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309237 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309228379 +0000 UTC m=+142.499757125 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309246 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/config: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309266 4872 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309296 4872 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309304 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309291401 +0000 UTC m=+142.499820017 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309246 4872 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309322 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309310842 +0000 UTC m=+142.499839568 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309336 4872 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309343 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309332982 +0000 UTC m=+142.499861738 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309360 4872 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309379 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309267 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309314 4872 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309413 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309396124 +0000 UTC m=+142.499924850 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309984 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309974212 +0000 UTC m=+142.500502838 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.309999 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.309990762 +0000 UTC m=+142.500519388 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310009 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert podName:fd040384-e861-44c9-bffc-1dc10acb8e89 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.310005113 +0000 UTC m=+142.500533729 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert") pod "apiserver-7bbb656c7d-7cglv" (UID: "fd040384-e861-44c9-bffc-1dc10acb8e89") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310024 4872 secret.go:188] Couldn't get secret openshift-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310033 4872 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.310051 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7"] Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310043 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.310038424 +0000 UTC m=+142.500567050 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync secret cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310139 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.310118326 +0000 UTC m=+142.500647032 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.310383 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle podName:6f555e5b-289d-40e7-9671-c47c69282416 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.310369434 +0000 UTC m=+142.500898330 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle") pod "apiserver-76f77b778f-rlfqc" (UID: "6f555e5b-289d-40e7-9671-c47c69282416") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.311376 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.311364794 +0000 UTC m=+142.501893410 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.327095 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.332874 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.337928 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcff0bb76_a914_4c38_90d2_f879b8c00b6e.slice/crio-8a3b816680f138f7feeeef5ffd4dd25fa563ef4059dd0cf74530001db06847ab WatchSource:0}: Error finding container 8a3b816680f138f7feeeef5ffd4dd25fa563ef4059dd0cf74530001db06847ab: Status 404 returned error can't find the container with id 8a3b816680f138f7feeeef5ffd4dd25fa563ef4059dd0cf74530001db06847ab Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.340260 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod847c2d27_b3a4_4dcb_a00e_307360b85f59.slice/crio-fbac56fe25806242f89ab860f70ab9fc5aa3b37eb9df4604c4c6c2d2130f5899 WatchSource:0}: Error finding container fbac56fe25806242f89ab860f70ab9fc5aa3b37eb9df4604c4c6c2d2130f5899: Status 404 returned error can't find the container with id fbac56fe25806242f89ab860f70ab9fc5aa3b37eb9df4604c4c6c2d2130f5899 Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.345391 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.364916 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.378503 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.388041 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.411111 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.426479 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.444075 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.445120 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.457806 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpzch\" (UniqueName: \"kubernetes.io/projected/fd040384-e861-44c9-bffc-1dc10acb8e89-kube-api-access-tpzch\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.467365 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.484324 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fb7cde4_b48a_4b5d_8352_54b49a017713.slice/crio-e4173ceb3f7d7696b8fdd622f0acbdaa933932bc65cb98da6147178c49a0edd0 WatchSource:0}: Error finding container e4173ceb3f7d7696b8fdd622f0acbdaa933932bc65cb98da6147178c49a0edd0: Status 404 returned error can't find the container with id e4173ceb3f7d7696b8fdd622f0acbdaa933932bc65cb98da6147178c49a0edd0 Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.486068 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.496778 4872 projected.go:288] Couldn't get configMap openshift-machine-api/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.496814 4872 projected.go:194] Error preparing data for projected volume kube-api-access-nvcrp for pod openshift-machine-api/machine-api-operator-5694c8668f-wcb54: failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.496873 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp podName:a1a5b7f5-7caa-4fe2-afb4-6c7176939e63 nodeName:}" failed. No retries permitted until 2025-10-09 08:19:43.996854947 +0000 UTC m=+142.187383563 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-nvcrp" (UniqueName: "kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp") pod "machine-api-operator-5694c8668f-wcb54" (UID: "a1a5b7f5-7caa-4fe2-afb4-6c7176939e63") : failed to sync configmap cache: timed out waiting for the condition Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.505442 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.526747 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.545989 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.564760 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.585623 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.594710 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.596064 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.597526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.597525 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr8vr\" (UniqueName: \"kubernetes.io/projected/6f555e5b-289d-40e7-9671-c47c69282416-kube-api-access-qr8vr\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.607437 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.610141 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ec987d5_2185_4e9e_96bc_307ea83ad0d0.slice/crio-b1f0e50645966169aba1d974a5eaef069d51835922ccf534b7572cc23fb4bbd2 WatchSource:0}: Error finding container b1f0e50645966169aba1d974a5eaef069d51835922ccf534b7572cc23fb4bbd2: Status 404 returned error can't find the container with id b1f0e50645966169aba1d974a5eaef069d51835922ccf534b7572cc23fb4bbd2 Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.636514 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.643314 4872 request.go:700] Waited for 1.627525677s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-oauth-apiserver/secrets?fieldSelector=metadata.name%3Detcd-client&limit=500&resourceVersion=0 Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.645821 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.646815 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4924a504_d8b5_406f_8379_19db356f0ba0.slice/crio-25a76cd42e61089d0b1c23c0330d9d1b20489efd21a442cd3fbb1b820a2ac781 WatchSource:0}: Error finding container 25a76cd42e61089d0b1c23c0330d9d1b20489efd21a442cd3fbb1b820a2ac781: Status 404 returned error can't find the container with id 25a76cd42e61089d0b1c23c0330d9d1b20489efd21a442cd3fbb1b820a2ac781 Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.652824 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda18c8409_0736_42ad_bb89_717d51c1c767.slice/crio-65aff83636984223f74a79339971275fa92149e82884c71fd5a949a95f9fe51d WatchSource:0}: Error finding container 65aff83636984223f74a79339971275fa92149e82884c71fd5a949a95f9fe51d: Status 404 returned error can't find the container with id 65aff83636984223f74a79339971275fa92149e82884c71fd5a949a95f9fe51d Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.669539 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.686560 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.708541 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.725811 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-86pkk"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.734865 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.747865 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-config\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.747941 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-key\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.747980 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748006 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748027 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748052 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/50f906df-d560-4cd6-9e57-0f92a754c74a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748077 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57z7n\" (UniqueName: \"kubernetes.io/projected/50f906df-d560-4cd6-9e57-0f92a754c74a-kube-api-access-57z7n\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748143 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edf09884-5c77-416e-b4ed-570657f567b6-serving-cert\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748188 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748213 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-cabundle\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748243 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-258cc\" (UniqueName: \"kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748291 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84fqd\" (UniqueName: \"kubernetes.io/projected/dce7a5e8-3074-463c-9ff5-b68ad42ced94-kube-api-access-84fqd\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748354 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748424 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w62xg\" (UniqueName: \"kubernetes.io/projected/edf09884-5c77-416e-b4ed-570657f567b6-kube-api-access-w62xg\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748458 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748485 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748577 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748603 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748622 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grnsg\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748661 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f314406-7129-4061-922a-7c2fba4019d9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748686 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748718 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg77g\" (UniqueName: \"kubernetes.io/projected/1f314406-7129-4061-922a-7c2fba4019d9-kube-api-access-fg77g\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.748762 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.750681 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z"] Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.751503 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.251486616 +0000 UTC m=+142.442015242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:43 crc kubenswrapper[4872]: W1009 08:19:43.762462 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36a1a915_0f81_4923_a700_98ca7584e702.slice/crio-912d837ede0d5d0aa2e00df9fdd0c8812fb2fe39a6fe583bccc7819cb12101cb WatchSource:0}: Error finding container 912d837ede0d5d0aa2e00df9fdd0c8812fb2fe39a6fe583bccc7819cb12101cb: Status 404 returned error can't find the container with id 912d837ede0d5d0aa2e00df9fdd0c8812fb2fe39a6fe583bccc7819cb12101cb Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.819771 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.849864 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850070 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850096 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850124 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850141 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850182 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grnsg\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850199 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f314406-7129-4061-922a-7c2fba4019d9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850215 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850240 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850255 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-csi-data-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850269 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850309 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg77g\" (UniqueName: \"kubernetes.io/projected/1f314406-7129-4061-922a-7c2fba4019d9-kube-api-access-fg77g\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850326 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-mountpoint-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850401 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-certs\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850442 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850476 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvjn9\" (UniqueName: \"kubernetes.io/projected/192a1e23-c791-475f-be33-26f4b383a383-kube-api-access-lvjn9\") pod \"downloads-7954f5f757-nm6xv\" (UID: \"192a1e23-c791-475f-be33-26f4b383a383\") " pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850527 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-socket-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.850543 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l4xt\" (UniqueName: \"kubernetes.io/projected/e729bc27-1f38-4b33-8c06-ad200e076cb9-kube-api-access-9l4xt\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.853204 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.353178515 +0000 UTC m=+142.543707191 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.855624 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856404 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b973560f-edd3-4939-9f0e-146834366a2b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-config\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856565 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-key\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856608 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856675 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856699 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856721 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856773 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/50f906df-d560-4cd6-9e57-0f92a754c74a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57z7n\" (UniqueName: \"kubernetes.io/projected/50f906df-d560-4cd6-9e57-0f92a754c74a-kube-api-access-57z7n\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856832 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dhs6\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-kube-api-access-5dhs6\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856902 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhbls\" (UniqueName: \"kubernetes.io/projected/83fcb79e-6159-4129-8260-5314af9a6e8e-kube-api-access-mhbls\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.856940 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edf09884-5c77-416e-b4ed-570657f567b6-serving-cert\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857031 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6e1d5c-4719-47d4-9faf-06d94241dca7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857075 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857098 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6e1d5c-4719-47d4-9faf-06d94241dca7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857120 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-cabundle\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857152 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-258cc\" (UniqueName: \"kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857186 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b973560f-edd3-4939-9f0e-146834366a2b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857234 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-node-bootstrap-token\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857288 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84fqd\" (UniqueName: \"kubernetes.io/projected/dce7a5e8-3074-463c-9ff5-b68ad42ced94-kube-api-access-84fqd\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857329 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80dfc75a-1353-4b3a-9561-bb5532034038-config-volume\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857523 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckp8r\" (UniqueName: \"kubernetes.io/projected/80dfc75a-1353-4b3a-9561-bb5532034038-kube-api-access-ckp8r\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857652 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80dfc75a-1353-4b3a-9561-bb5532034038-metrics-tls\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w62xg\" (UniqueName: \"kubernetes.io/projected/edf09884-5c77-416e-b4ed-570657f567b6-kube-api-access-w62xg\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857719 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6e1d5c-4719-47d4-9faf-06d94241dca7-config\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857743 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857767 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-registration-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857841 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857892 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.857971 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.864219 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-config\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.865596 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.365580083 +0000 UTC m=+142.556108789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.865587 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.866385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/1f314406-7129-4061-922a-7c2fba4019d9-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.866406 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/edf09884-5c77-416e-b4ed-570657f567b6-service-ca-bundle\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.866835 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-plugins-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.866863 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snct7\" (UniqueName: \"kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.866987 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.867026 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.867055 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-key\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.868125 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/50f906df-d560-4cd6-9e57-0f92a754c74a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.868912 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/dce7a5e8-3074-463c-9ff5-b68ad42ced94-signing-cabundle\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.869689 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.877441 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.878354 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.879338 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/edf09884-5c77-416e-b4ed-570657f567b6-serving-cert\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.888403 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.900386 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.900538 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-bkhr6"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.909432 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grnsg\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.919879 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.930312 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.954914 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg77g\" (UniqueName: \"kubernetes.io/projected/1f314406-7129-4061-922a-7c2fba4019d9-kube-api-access-fg77g\") pod \"control-plane-machine-set-operator-78cbb6b69f-b5qvz\" (UID: \"1f314406-7129-4061-922a-7c2fba4019d9\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.956783 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968086 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.968343 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.468314573 +0000 UTC m=+142.658843199 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6e1d5c-4719-47d4-9faf-06d94241dca7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968614 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b973560f-edd3-4939-9f0e-146834366a2b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968635 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-node-bootstrap-token\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968725 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80dfc75a-1353-4b3a-9561-bb5532034038-config-volume\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968755 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968781 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckp8r\" (UniqueName: \"kubernetes.io/projected/80dfc75a-1353-4b3a-9561-bb5532034038-kube-api-access-ckp8r\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968775 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968800 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80dfc75a-1353-4b3a-9561-bb5532034038-metrics-tls\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968922 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6e1d5c-4719-47d4-9faf-06d94241dca7-config\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.968953 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-registration-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969003 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969081 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snct7\" (UniqueName: \"kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969165 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-plugins-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969224 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969255 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969332 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-csi-data-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969355 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969412 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-mountpoint-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-certs\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969499 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvjn9\" (UniqueName: \"kubernetes.io/projected/192a1e23-c791-475f-be33-26f4b383a383-kube-api-access-lvjn9\") pod \"downloads-7954f5f757-nm6xv\" (UID: \"192a1e23-c791-475f-be33-26f4b383a383\") " pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969551 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-socket-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.969573 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l4xt\" (UniqueName: \"kubernetes.io/projected/e729bc27-1f38-4b33-8c06-ad200e076cb9-kube-api-access-9l4xt\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.974143 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84fqd\" (UniqueName: \"kubernetes.io/projected/dce7a5e8-3074-463c-9ff5-b68ad42ced94-kube-api-access-84fqd\") pod \"service-ca-9c57cc56f-fgzt2\" (UID: \"dce7a5e8-3074-463c-9ff5-b68ad42ced94\") " pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.979823 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-mountpoint-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.980335 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4c6e1d5c-4719-47d4-9faf-06d94241dca7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.981178 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c6e1d5c-4719-47d4-9faf-06d94241dca7-config\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.981363 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-registration-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.981621 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.982379 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.982439 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-csi-data-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.982482 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-socket-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: E1009 08:19:43.990377 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.490356465 +0000 UTC m=+142.680885091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.990920 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/83fcb79e-6159-4129-8260-5314af9a6e8e-plugins-dir\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.990971 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/80dfc75a-1353-4b3a-9561-bb5532034038-config-volume\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.991482 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b973560f-edd3-4939-9f0e-146834366a2b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.991594 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.991654 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dhs6\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-kube-api-access-5dhs6\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.991705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhbls\" (UniqueName: \"kubernetes.io/projected/83fcb79e-6159-4129-8260-5314af9a6e8e-kube-api-access-mhbls\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.991788 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6e1d5c-4719-47d4-9faf-06d94241dca7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.992547 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:43 crc kubenswrapper[4872]: I1009 08:19:43.993719 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b973560f-edd3-4939-9f0e-146834366a2b-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:43.999435 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.003833 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.007820 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-node-bootstrap-token\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.009112 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/80dfc75a-1353-4b3a-9561-bb5532034038-metrics-tls\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.010061 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.010431 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/b973560f-edd3-4939-9f0e-146834366a2b-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.010491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/e729bc27-1f38-4b33-8c06-ad200e076cb9-certs\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.012766 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q"] Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.014936 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w62xg\" (UniqueName: \"kubernetes.io/projected/edf09884-5c77-416e-b4ed-570657f567b6-kube-api-access-w62xg\") pod \"authentication-operator-69f744f599-dk2r4\" (UID: \"edf09884-5c77-416e-b4ed-570657f567b6\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.016885 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57z7n\" (UniqueName: \"kubernetes.io/projected/50f906df-d560-4cd6-9e57-0f92a754c74a-kube-api-access-57z7n\") pod \"package-server-manager-789f6589d5-2775d\" (UID: \"50f906df-d560-4cd6-9e57-0f92a754c74a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:44 crc kubenswrapper[4872]: W1009 08:19:44.019052 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb6c618e_f78b_40eb_8556_3824be52910d.slice/crio-ba3ac18e9e1077d3de32afa8447c799e42f13585558d090c08fa1e337c4d4818 WatchSource:0}: Error finding container ba3ac18e9e1077d3de32afa8447c799e42f13585558d090c08fa1e337c4d4818: Status 404 returned error can't find the container with id ba3ac18e9e1077d3de32afa8447c799e42f13585558d090c08fa1e337c4d4818 Oct 09 08:19:44 crc kubenswrapper[4872]: W1009 08:19:44.022540 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod75af6421_130b_4287_a014_d17512f708e4.slice/crio-9d954d2c5cde8f3daf6ede45b7282d0ed76e396d6a4a86f5596d3bf6710ed3c6 WatchSource:0}: Error finding container 9d954d2c5cde8f3daf6ede45b7282d0ed76e396d6a4a86f5596d3bf6710ed3c6: Status 404 returned error can't find the container with id 9d954d2c5cde8f3daf6ede45b7282d0ed76e396d6a4a86f5596d3bf6710ed3c6 Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.025028 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-258cc\" (UniqueName: \"kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc\") pod \"collect-profiles-29333295-tptvl\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.065173 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckp8r\" (UniqueName: \"kubernetes.io/projected/80dfc75a-1353-4b3a-9561-bb5532034038-kube-api-access-ckp8r\") pod \"dns-default-dfq2v\" (UID: \"80dfc75a-1353-4b3a-9561-bb5532034038\") " pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.093031 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.093127 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.593111556 +0000 UTC m=+142.783640182 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.093572 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvcrp\" (UniqueName: \"kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.094181 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.094795 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.594777827 +0000 UTC m=+142.785306453 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.095333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snct7\" (UniqueName: \"kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7\") pod \"console-f9d7485db-4gkkd\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.096298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.101292 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvcrp\" (UniqueName: \"kubernetes.io/projected/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-kube-api-access-nvcrp\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.101545 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.132698 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvjn9\" (UniqueName: \"kubernetes.io/projected/192a1e23-c791-475f-be33-26f4b383a383-kube-api-access-lvjn9\") pod \"downloads-7954f5f757-nm6xv\" (UID: \"192a1e23-c791-475f-be33-26f4b383a383\") " pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.132298 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-dt452"] Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.155708 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l4xt\" (UniqueName: \"kubernetes.io/projected/e729bc27-1f38-4b33-8c06-ad200e076cb9-kube-api-access-9l4xt\") pod \"machine-config-server-9dqkl\" (UID: \"e729bc27-1f38-4b33-8c06-ad200e076cb9\") " pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.162087 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.169198 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4c6e1d5c-4719-47d4-9faf-06d94241dca7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-7f488\" (UID: \"4c6e1d5c-4719-47d4-9faf-06d94241dca7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.170373 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dhs6\" (UniqueName: \"kubernetes.io/projected/b973560f-edd3-4939-9f0e-146834366a2b-kube-api-access-5dhs6\") pod \"cluster-image-registry-operator-dc59b4c8b-6k8nz\" (UID: \"b973560f-edd3-4939-9f0e-146834366a2b\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.172896 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:44 crc kubenswrapper[4872]: W1009 08:19:44.180587 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7af426f_a554_434f_9735_922bebb651bf.slice/crio-9ed8852f791ee3068bd45f4a157c4b34d76f83a7207680f28036770bbe815334 WatchSource:0}: Error finding container 9ed8852f791ee3068bd45f4a157c4b34d76f83a7207680f28036770bbe815334: Status 404 returned error can't find the container with id 9ed8852f791ee3068bd45f4a157c4b34d76f83a7207680f28036770bbe815334 Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.187238 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhbls\" (UniqueName: \"kubernetes.io/projected/83fcb79e-6159-4129-8260-5314af9a6e8e-kube-api-access-mhbls\") pod \"csi-hostpathplugin-fbd5f\" (UID: \"83fcb79e-6159-4129-8260-5314af9a6e8e\") " pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.196169 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.196584 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.696568029 +0000 UTC m=+142.887096655 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.250969 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.280730 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.287994 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" event={"ID":"4924a504-d8b5-406f-8379-19db356f0ba0","Type":"ContainerStarted","Data":"fb3148b442dce873d75dd2975302082e9c79b8bb5e4909d90abc85d20b62f8b3"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.288039 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" event={"ID":"4924a504-d8b5-406f-8379-19db356f0ba0","Type":"ContainerStarted","Data":"25a76cd42e61089d0b1c23c0330d9d1b20489efd21a442cd3fbb1b820a2ac781"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.288338 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.291453 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" event={"ID":"847c2d27-b3a4-4dcb-a00e-307360b85f59","Type":"ContainerStarted","Data":"fbac56fe25806242f89ab860f70ab9fc5aa3b37eb9df4604c4c6c2d2130f5899"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.297864 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.298268 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.798254257 +0000 UTC m=+142.988782893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.299031 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.299440 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.302932 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.303294 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" event={"ID":"982bf7fe-335b-4fb6-be38-62de01442760","Type":"ContainerStarted","Data":"3680b0734f4a4dbc89283cbc672d6e97f2506128c7494431d365ea783f56f376"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.303325 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" event={"ID":"982bf7fe-335b-4fb6-be38-62de01442760","Type":"ContainerStarted","Data":"d48a8fb931ee4986b4f14a41e469e57ee557e8b52b3d678b10a76630ead1a854"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.304410 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.304622 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-9dqkl" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.307455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" event={"ID":"75af6421-130b-4287-a014-d17512f708e4","Type":"ContainerStarted","Data":"9d954d2c5cde8f3daf6ede45b7282d0ed76e396d6a4a86f5596d3bf6710ed3c6"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.308689 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" event={"ID":"3be12f5f-7fdb-4d42-b66e-42bccc2980e7","Type":"ContainerStarted","Data":"ac1dbdcd3366b4b0ced75888c189510ed6dc50aeeb4c7359401e240e90c71353"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.309980 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" event={"ID":"f9b56021-cd13-4c6a-9f73-41db5b77d85b","Type":"ContainerStarted","Data":"17457e12fa2e31e824c332d0f3f85082786a4b930f0a0f557944263981921d65"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.312445 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.313543 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" event={"ID":"44e52fef-7688-437b-8c44-af285a46273d","Type":"ContainerStarted","Data":"92c2b5410ca8c63a3c2700d8ed26fc178d2ce9f43d0fff7e3be09614845dfccf"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.313569 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" event={"ID":"44e52fef-7688-437b-8c44-af285a46273d","Type":"ContainerStarted","Data":"04451fe87e200e689a8c742c6b6acdc6a01b7875a61f3e689dd812c1b634bb05"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.315487 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" event={"ID":"cff0bb76-a914-4c38-90d2-f879b8c00b6e","Type":"ContainerStarted","Data":"7ba82edac24a54bea777af4ce07fb7b521bc2059c2a61faa2b6da04895f24b1f"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.315515 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" event={"ID":"cff0bb76-a914-4c38-90d2-f879b8c00b6e","Type":"ContainerStarted","Data":"8a3b816680f138f7feeeef5ffd4dd25fa563ef4059dd0cf74530001db06847ab"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.316253 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.317180 4872 patch_prober.go:28] interesting pod/console-operator-58897d9998-bvcfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.317219 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" podUID="cff0bb76-a914-4c38-90d2-f879b8c00b6e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": dial tcp 10.217.0.10:8443: connect: connection refused" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.318391 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" event={"ID":"81d1c103-d6b5-4e05-a262-10a658db5948","Type":"ContainerStarted","Data":"1d04b5af58e1dcf5edd5ef82122c0870ef9c54a44bfbbaa601896799b41a0f08"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.331761 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" event={"ID":"035a7dd6-e488-4af1-ae0f-345f6232751f","Type":"ContainerStarted","Data":"f2fa58a6c271e63bc1f809adbc9890bda3796e085f1a030aec4d600ea397276a"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.337651 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.343261 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gsqtq" event={"ID":"61adb101-0d3a-4018-aa4f-4ed4a435148b","Type":"ContainerStarted","Data":"badb341c28c80bed1d513b81fa088f8c4e83d06843cf2fb6fad14f7390922ee1"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.343295 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gsqtq" event={"ID":"61adb101-0d3a-4018-aa4f-4ed4a435148b","Type":"ContainerStarted","Data":"ea53dfa7fb473823ca83a428402012409c92d2fe5650c9c555932610f9581cf9"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.355632 4872 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-978jd container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" start-of-body= Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.355870 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" podUID="982bf7fe-335b-4fb6-be38-62de01442760" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.23:8443/healthz\": dial tcp 10.217.0.23:8443: connect: connection refused" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.394955 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" event={"ID":"1b2002e1-bd93-447f-9a8b-e647308be981","Type":"ContainerStarted","Data":"f379eada665df2ae77872e2d51a441cd20c5b592b782f26e8ec881ea9569b9b8"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399234 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399410 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399461 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399491 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399512 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399538 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399558 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399582 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399627 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399782 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399810 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399834 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399882 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.399943 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.400027 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.400790 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:44.900770011 +0000 UTC m=+143.091298637 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.402289 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-config\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.407304 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-audit-policies\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.407708 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-trusted-ca-bundle\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.408238 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-etcd-serving-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.408508 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.408586 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-etcd-client\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.409313 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-image-import-ca\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.409355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd040384-e861-44c9-bffc-1dc10acb8e89-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.409394 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-images\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.419999 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fd040384-e861-44c9-bffc-1dc10acb8e89-serving-cert\") pod \"apiserver-7bbb656c7d-7cglv\" (UID: \"fd040384-e861-44c9-bffc-1dc10acb8e89\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.421429 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/6f555e5b-289d-40e7-9671-c47c69282416-audit\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.421798 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-encryption-config\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.429180 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-serving-cert\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.430123 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a1a5b7f5-7caa-4fe2-afb4-6c7176939e63-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-wcb54\" (UID: \"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.438449 4872 generic.go:334] "Generic (PLEG): container finished" podID="8a5e4235-b5ba-495b-8462-8796c968d1f4" containerID="1093adde2404fae060231b84c497b5eb945a835a9f99e604f8197971f00224c6" exitCode=0 Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.438530 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" event={"ID":"8a5e4235-b5ba-495b-8462-8796c968d1f4","Type":"ContainerDied","Data":"1093adde2404fae060231b84c497b5eb945a835a9f99e604f8197971f00224c6"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.446180 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f555e5b-289d-40e7-9671-c47c69282416-etcd-client\") pod \"apiserver-76f77b778f-rlfqc\" (UID: \"6f555e5b-289d-40e7-9671-c47c69282416\") " pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.453333 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.497870 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.519977 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.526425 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.530959 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.030942768 +0000 UTC m=+143.221471474 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.544987 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" event={"ID":"bb6c618e-f78b-40eb-8556-3824be52910d","Type":"ContainerStarted","Data":"ba3ac18e9e1077d3de32afa8447c799e42f13585558d090c08fa1e337c4d4818"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.571471 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-86pkk" event={"ID":"36a1a915-0f81-4923-a700-98ca7584e702","Type":"ContainerStarted","Data":"912d837ede0d5d0aa2e00df9fdd0c8812fb2fe39a6fe583bccc7819cb12101cb"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.583296 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" event={"ID":"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6","Type":"ContainerStarted","Data":"451697cd8e9b0e449651e9be6e3e1e38505e5888c2f32c282b8a1aa57c86790a"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.599043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" event={"ID":"4fb7cde4-b48a-4b5d-8352-54b49a017713","Type":"ContainerStarted","Data":"4dc18aeaea90dff0cb4455766e7ea8b7bad2bfb384dd8947de802beb61d569a8"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.599117 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" event={"ID":"4fb7cde4-b48a-4b5d-8352-54b49a017713","Type":"ContainerStarted","Data":"e4173ceb3f7d7696b8fdd622f0acbdaa933932bc65cb98da6147178c49a0edd0"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.620367 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d"] Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.626058 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" event={"ID":"a18c8409-0736-42ad-bb89-717d51c1c767","Type":"ContainerStarted","Data":"65aff83636984223f74a79339971275fa92149e82884c71fd5a949a95f9fe51d"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.627459 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.627717 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.628198 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.128179721 +0000 UTC m=+143.318708357 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.637862 4872 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xlstb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.637909 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.639924 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" event={"ID":"03a6afc5-c556-4b61-a41a-c6004edc386c","Type":"ContainerStarted","Data":"594c407a37729f8a9cf7abcb66b5095f6c67772c7909bcf60b86908d5c37e726"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.640825 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.643213 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" event={"ID":"1ec987d5-2185-4e9e-96bc-307ea83ad0d0","Type":"ContainerStarted","Data":"d48c8e240fba203159cb8d2e94ca064cbf1c6a9984f15af69f3788f57e453e30"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.643278 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" event={"ID":"1ec987d5-2185-4e9e-96bc-307ea83ad0d0","Type":"ContainerStarted","Data":"b1f0e50645966169aba1d974a5eaef069d51835922ccf534b7572cc23fb4bbd2"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.645057 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" event={"ID":"159cd083-1078-4cb3-9142-b985c7a53823","Type":"ContainerStarted","Data":"65a261fd1868426df4a0112a4d2357a866f3fcadd42162a766737ab182b5bba1"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.648047 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" event={"ID":"11d63870-f531-4733-a134-6de125f2a7a5","Type":"ContainerStarted","Data":"1c827d4050857d51f190eb163bc5d492fe0a421dc9d14d024a32d7ae700145d9"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.658723 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" event={"ID":"e7af426f-a554-434f-9735-922bebb651bf","Type":"ContainerStarted","Data":"9ed8852f791ee3068bd45f4a157c4b34d76f83a7207680f28036770bbe815334"} Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.661174 4872 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h2ll5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.661207 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" podUID="03a6afc5-c556-4b61-a41a-c6004edc386c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.666302 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.682577 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.685029 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-5lh66" podStartSLOduration=123.685009653 podStartE2EDuration="2m3.685009653s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:44.683879099 +0000 UTC m=+142.874407745" watchObservedRunningTime="2025-10-09 08:19:44.685009653 +0000 UTC m=+142.875538279" Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.731443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.751788 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.251769238 +0000 UTC m=+143.442297864 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.833347 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.840020 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.339992946 +0000 UTC m=+143.530521562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.840116 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.841762 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.34175403 +0000 UTC m=+143.532282656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.880494 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" podStartSLOduration=123.88047873 podStartE2EDuration="2m3.88047873s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:44.846464693 +0000 UTC m=+143.036993339" watchObservedRunningTime="2025-10-09 08:19:44.88047873 +0000 UTC m=+143.071007356" Oct 09 08:19:44 crc kubenswrapper[4872]: W1009 08:19:44.896983 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50f906df_d560_4cd6_9e57_0f92a754c74a.slice/crio-1a1ba9052f5e6ffb8d358b489bd3558153a83deef556b220b3614b40f6c20220 WatchSource:0}: Error finding container 1a1ba9052f5e6ffb8d358b489bd3558153a83deef556b220b3614b40f6c20220: Status 404 returned error can't find the container with id 1a1ba9052f5e6ffb8d358b489bd3558153a83deef556b220b3614b40f6c20220 Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.897382 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl"] Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.933412 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-nm6xv"] Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.942374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.944510 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.444468 +0000 UTC m=+143.634996626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.946515 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:44 crc kubenswrapper[4872]: E1009 08:19:44.947054 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.447039738 +0000 UTC m=+143.637568364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:44 crc kubenswrapper[4872]: I1009 08:19:44.968386 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz"] Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.038389 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-dfq2v"] Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.048096 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.048531 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.54850604 +0000 UTC m=+143.739034666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.059311 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488"] Oct 09 08:19:45 crc kubenswrapper[4872]: W1009 08:19:45.068300 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod192a1e23_c791_475f_be33_26f4b383a383.slice/crio-254fe275296d67f146a8cb57e6257ab11474b22a878d6b917053e0557aca7d17 WatchSource:0}: Error finding container 254fe275296d67f146a8cb57e6257ab11474b22a878d6b917053e0557aca7d17: Status 404 returned error can't find the container with id 254fe275296d67f146a8cb57e6257ab11474b22a878d6b917053e0557aca7d17 Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.119365 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" podStartSLOduration=123.119348179 podStartE2EDuration="2m3.119348179s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:45.117223424 +0000 UTC m=+143.307752060" watchObservedRunningTime="2025-10-09 08:19:45.119348179 +0000 UTC m=+143.309876805" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.149603 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.149954 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.649941671 +0000 UTC m=+143.840470297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.214600 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.258866 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.259396 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.759367626 +0000 UTC m=+143.949896252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.259439 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.259846 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.75983186 +0000 UTC m=+143.950360566 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.364655 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.364720 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.366201 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.366610 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.866591033 +0000 UTC m=+144.057119669 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.470519 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.471238 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:45.971224122 +0000 UTC m=+144.161752758 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.581301 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-fgzt2"] Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.584741 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.585062 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.085044311 +0000 UTC m=+144.275572937 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.686615 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.687332 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.187317667 +0000 UTC m=+144.377846293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.693423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" event={"ID":"1f314406-7129-4061-922a-7c2fba4019d9","Type":"ContainerStarted","Data":"cb220d839885cf4d90aa5f6e73b45c51c6e8eae6ea3e213e97de776b73f49fa3"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.710528 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" event={"ID":"4c6e1d5c-4719-47d4-9faf-06d94241dca7","Type":"ContainerStarted","Data":"b6fc0d025b1e589304f1cd6aa687652540eaf334cae4afedb68e72d2d7d2304e"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.787931 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.788396 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.288381597 +0000 UTC m=+144.478910223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.794328 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" event={"ID":"f9b56021-cd13-4c6a-9f73-41db5b77d85b","Type":"ContainerStarted","Data":"d7ec0adb2e4e26e1acc8c8415ecc7c526ca3af82e499b1424ee29ede8cf9e2dc"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.795509 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.797377 4872 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-mqrzg container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" start-of-body= Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.797451 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" podUID="f9b56021-cd13-4c6a-9f73-41db5b77d85b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.41:8443/healthz\": dial tcp 10.217.0.41:8443: connect: connection refused" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.853576 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" podStartSLOduration=124.853562633 podStartE2EDuration="2m4.853562633s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:45.823028303 +0000 UTC m=+144.013556949" watchObservedRunningTime="2025-10-09 08:19:45.853562633 +0000 UTC m=+144.044091259" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.855880 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9dqkl" event={"ID":"e729bc27-1f38-4b33-8c06-ad200e076cb9","Type":"ContainerStarted","Data":"2d97b031214fbba9fb8efc3121d7fecedf4e9fbd7a18472e837af77aa379d7fb"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.860114 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" podStartSLOduration=123.860095502 podStartE2EDuration="2m3.860095502s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:45.8531295 +0000 UTC m=+144.043658136" watchObservedRunningTime="2025-10-09 08:19:45.860095502 +0000 UTC m=+144.050624128" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.890969 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.891346 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.391330754 +0000 UTC m=+144.581859380 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.915773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" event={"ID":"847c2d27-b3a4-4dcb-a00e-307360b85f59","Type":"ContainerStarted","Data":"97ab5ba7ce862c0c7392c8abf9aa27780cd7a271090df9d3394ad1a7c3378d4d"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.922581 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" podStartSLOduration=123.922565336 podStartE2EDuration="2m3.922565336s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:45.922119533 +0000 UTC m=+144.112648159" watchObservedRunningTime="2025-10-09 08:19:45.922565336 +0000 UTC m=+144.113093962" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.945015 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" event={"ID":"7b2f7d7a-26a5-4566-916c-cf5af4d0cec6","Type":"ContainerStarted","Data":"b8e9f4689f682ca0f7a0603a4fd682f0ab49322c2605b0c869dd1925d0022667"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.961129 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-2lf9z" podStartSLOduration=123.961112771 podStartE2EDuration="2m3.961112771s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:45.959275225 +0000 UTC m=+144.149803851" watchObservedRunningTime="2025-10-09 08:19:45.961112771 +0000 UTC m=+144.151641397" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.981354 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" event={"ID":"a18c8409-0736-42ad-bb89-717d51c1c767","Type":"ContainerStarted","Data":"2e2d171e335a67b9100fa1dab0e810e1e661ef0b2ca2f70f5ec37eec6a35da65"} Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.983911 4872 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xlstb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.983977 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 09 08:19:45 crc kubenswrapper[4872]: I1009 08:19:45.991879 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:45 crc kubenswrapper[4872]: E1009 08:19:45.993134 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.493116026 +0000 UTC m=+144.683644662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.015136 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gsqtq" podStartSLOduration=125.015113306 podStartE2EDuration="2m5.015113306s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.014314892 +0000 UTC m=+144.204843518" watchObservedRunningTime="2025-10-09 08:19:46.015113306 +0000 UTC m=+144.205641932" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.036202 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" event={"ID":"3be12f5f-7fdb-4d42-b66e-42bccc2980e7","Type":"ContainerStarted","Data":"16de11db263bec15ee5b3f1fbd4a33723cefde2d929ee55169092604b2420206"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.056933 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfq2v" event={"ID":"80dfc75a-1353-4b3a-9561-bb5532034038","Type":"ContainerStarted","Data":"6450ad1400a85d9a0accf081217a09d4abe28c0b563c11a31d65c70348cee9e1"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.065030 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" event={"ID":"81d1c103-d6b5-4e05-a262-10a658db5948","Type":"ContainerStarted","Data":"ce7835b471b123ab8bd2e51a6b47370b9c84b8fea81d7e778e89592e85dd492d"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.066664 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" podStartSLOduration=124.066604886 podStartE2EDuration="2m4.066604886s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.049993249 +0000 UTC m=+144.240521885" watchObservedRunningTime="2025-10-09 08:19:46.066604886 +0000 UTC m=+144.257133512" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.085417 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" podStartSLOduration=124.085397108 podStartE2EDuration="2m4.085397108s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.083023536 +0000 UTC m=+144.273552192" watchObservedRunningTime="2025-10-09 08:19:46.085397108 +0000 UTC m=+144.275925734" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.091177 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" event={"ID":"50f906df-d560-4cd6-9e57-0f92a754c74a","Type":"ContainerStarted","Data":"1a1ba9052f5e6ffb8d358b489bd3558153a83deef556b220b3614b40f6c20220"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.094669 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.097070 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.597052493 +0000 UTC m=+144.787581209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.119577 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" podStartSLOduration=125.119550469 podStartE2EDuration="2m5.119550469s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.107723389 +0000 UTC m=+144.298252015" watchObservedRunningTime="2025-10-09 08:19:46.119550469 +0000 UTC m=+144.310079105" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.121732 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" event={"ID":"035a7dd6-e488-4af1-ae0f-345f6232751f","Type":"ContainerStarted","Data":"6a392468ed14ef7c91f07501c42b0ae0eed963b0e57a71d3abd03cf3f78237f6"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.169065 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-28xtq" event={"ID":"4fb7cde4-b48a-4b5d-8352-54b49a017713","Type":"ContainerStarted","Data":"fd1c1cae9af7121951cf50d796e2df2b8cc526a41d5ffd1c7be7b273ee8e715c"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.180549 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-pn8h7" podStartSLOduration=125.180514007 podStartE2EDuration="2m5.180514007s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.178835776 +0000 UTC m=+144.369364412" watchObservedRunningTime="2025-10-09 08:19:46.180514007 +0000 UTC m=+144.371042633" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.197505 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.197912 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.697887366 +0000 UTC m=+144.888415992 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.198035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.200731 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.700715212 +0000 UTC m=+144.891243838 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.203929 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.214554 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-86pkk" event={"ID":"36a1a915-0f81-4923-a700-98ca7584e702","Type":"ContainerStarted","Data":"6e19e70706042be4fb6e5f06cc46c4263be20b221b0eece2ba897be551a98177"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.221342 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" event={"ID":"b03f0070-47ed-455e-b4b8-7a86fa5818af","Type":"ContainerStarted","Data":"22dceb6618e8d91e18edbf7f8f0f104a78fc4d01b2ceb1b5be2bf7ada304cc49"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.226928 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-54wdx" podStartSLOduration=125.22689794 podStartE2EDuration="2m5.22689794s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.214350538 +0000 UTC m=+144.404879184" watchObservedRunningTime="2025-10-09 08:19:46.22689794 +0000 UTC m=+144.417426576" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.269281 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" event={"ID":"e7af426f-a554-434f-9735-922bebb651bf","Type":"ContainerStarted","Data":"01b1fba70d85645e8695dc74f1e15f2d4d93a8d17a6513ec4ac631124940b23d"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.272425 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:46 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:46 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:46 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.272483 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.273629 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" event={"ID":"1b2002e1-bd93-447f-9a8b-e647308be981","Type":"ContainerStarted","Data":"00d624bb923602f78c5d8cbf3071ac8a881571f4617026555c45120bfda241b3"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.284017 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.290625 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" podStartSLOduration=124.290600802 podStartE2EDuration="2m4.290600802s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.261787894 +0000 UTC m=+144.452316540" watchObservedRunningTime="2025-10-09 08:19:46.290600802 +0000 UTC m=+144.481129428" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.301017 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.302149 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.802128133 +0000 UTC m=+144.992656759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.302218 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-86pkk" podStartSLOduration=6.302196735 podStartE2EDuration="6.302196735s" podCreationTimestamp="2025-10-09 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.28202754 +0000 UTC m=+144.472556166" watchObservedRunningTime="2025-10-09 08:19:46.302196735 +0000 UTC m=+144.492725371" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.311936 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-mqcvd" podStartSLOduration=125.311918411 podStartE2EDuration="2m5.311918411s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.308974192 +0000 UTC m=+144.499502828" watchObservedRunningTime="2025-10-09 08:19:46.311918411 +0000 UTC m=+144.502447027" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.377114 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-fbd5f"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.385516 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" event={"ID":"8a5e4235-b5ba-495b-8462-8796c968d1f4","Type":"ContainerStarted","Data":"12f19027bf66516f26a4ac75aa875f9fae50bbdb81033509be0dcaeddf7cf4ac"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.386434 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.402584 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.404095 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:46.90407488 +0000 UTC m=+145.094603586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.419009 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" event={"ID":"03a6afc5-c556-4b61-a41a-c6004edc386c","Type":"ContainerStarted","Data":"2b5aad87b9cc553ba497b069f37935fc3845030d5f4bae8d1bc2516faea56e5d"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.420055 4872 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h2ll5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.420095 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" podUID="03a6afc5-c556-4b61-a41a-c6004edc386c" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.432116 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" event={"ID":"1ec987d5-2185-4e9e-96bc-307ea83ad0d0","Type":"ContainerStarted","Data":"a0d882f7fedb5d7c7736add81bae0e414055f2dfd8f25bb85ab12b937c97c811"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.448074 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-lm8n7" event={"ID":"159cd083-1078-4cb3-9142-b985c7a53823","Type":"ContainerStarted","Data":"87817fdc5ee9e84cd767b54cd163db6fc36fb3ce11ea63ba1810d9b5f733ee1f"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.455941 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-wcb54"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.475115 4872 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-g747n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" start-of-body= Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.475248 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": dial tcp 10.217.0.15:6443: connect: connection refused" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.477008 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.477089 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dk2r4"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.477160 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" event={"ID":"75af6421-130b-4287-a014-d17512f708e4","Type":"ContainerStarted","Data":"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.477222 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.477278 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-rlfqc"] Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.491908 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" event={"ID":"bb6c618e-f78b-40eb-8556-3824be52910d","Type":"ContainerStarted","Data":"95c68ac04b4b6f3bf899b76b4ef96c460d7fac7d005ecf7dc351d0255bc84997"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.494121 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" event={"ID":"11d63870-f531-4733-a134-6de125f2a7a5","Type":"ContainerStarted","Data":"adcb298d92f18e17e08a30a0baff90507c2a43522e32a29a3389715a3e18fe97"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.504434 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.504529 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.00450172 +0000 UTC m=+145.195030346 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.504776 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.504969 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" event={"ID":"44e52fef-7688-437b-8c44-af285a46273d","Type":"ContainerStarted","Data":"66cb363df09e477b31f2d614fa4463245e2b9b777f6d2e27aeb48285caa24ae8"} Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.506661 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.006652575 +0000 UTC m=+145.197181201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.539275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nm6xv" event={"ID":"192a1e23-c791-475f-be33-26f4b383a383","Type":"ContainerStarted","Data":"254fe275296d67f146a8cb57e6257ab11474b22a878d6b917053e0557aca7d17"} Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.539324 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.546704 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.546765 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.564827 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-978jd" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.605971 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.607446 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.107420656 +0000 UTC m=+145.297949332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.614597 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-l275r" podStartSLOduration=125.614532523 podStartE2EDuration="2m5.614532523s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.613919314 +0000 UTC m=+144.804447940" watchObservedRunningTime="2025-10-09 08:19:46.614532523 +0000 UTC m=+144.805061149" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.615735 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" podStartSLOduration=125.615729829 podStartE2EDuration="2m5.615729829s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.581691182 +0000 UTC m=+144.772219828" watchObservedRunningTime="2025-10-09 08:19:46.615729829 +0000 UTC m=+144.806258455" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.707961 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.708584 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.208569099 +0000 UTC m=+145.399097725 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.738062 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" podStartSLOduration=125.738047877 podStartE2EDuration="2m5.738047877s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.735570652 +0000 UTC m=+144.926099278" watchObservedRunningTime="2025-10-09 08:19:46.738047877 +0000 UTC m=+144.928576503" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.763172 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fv2jp" podStartSLOduration=125.763153832 podStartE2EDuration="2m5.763153832s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.648386245 +0000 UTC m=+144.838914881" watchObservedRunningTime="2025-10-09 08:19:46.763153832 +0000 UTC m=+144.953682458" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.819254 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.819477 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.319445087 +0000 UTC m=+145.509973723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.819975 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.820297 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.320279903 +0000 UTC m=+145.510808529 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.821907 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" podStartSLOduration=125.821892412 podStartE2EDuration="2m5.821892412s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.775143167 +0000 UTC m=+144.965671793" watchObservedRunningTime="2025-10-09 08:19:46.821892412 +0000 UTC m=+145.012421038" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.824958 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" podStartSLOduration=125.824941575 podStartE2EDuration="2m5.824941575s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.820281773 +0000 UTC m=+145.010810409" watchObservedRunningTime="2025-10-09 08:19:46.824941575 +0000 UTC m=+145.015470221" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.884688 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-mlfjj" podStartSLOduration=125.884670515 podStartE2EDuration="2m5.884670515s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:46.884115988 +0000 UTC m=+145.074644624" watchObservedRunningTime="2025-10-09 08:19:46.884670515 +0000 UTC m=+145.075199141" Oct 09 08:19:46 crc kubenswrapper[4872]: I1009 08:19:46.929387 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:46 crc kubenswrapper[4872]: E1009 08:19:46.929855 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.429836022 +0000 UTC m=+145.620364658 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.003135 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-nm6xv" podStartSLOduration=126.003120385 podStartE2EDuration="2m6.003120385s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.001430903 +0000 UTC m=+145.191959529" watchObservedRunningTime="2025-10-09 08:19:47.003120385 +0000 UTC m=+145.193649021" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.031904 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.032548 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.532534131 +0000 UTC m=+145.723062757 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.137726 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.137976 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.637962494 +0000 UTC m=+145.828491110 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.228334 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:47 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:47 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:47 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.228400 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.239426 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.239746 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.739735305 +0000 UTC m=+145.930263931 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.340649 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.340919 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.840888648 +0000 UTC m=+146.031417284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.341558 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.341962 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.84195229 +0000 UTC m=+146.032480916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.442358 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.442588 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.942553816 +0000 UTC m=+146.133082442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.442653 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.443083 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:47.943066202 +0000 UTC m=+146.133594828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.452739 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.453777 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.456461 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.479533 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.540349 4872 patch_prober.go:28] interesting pod/console-operator-58897d9998-bvcfj container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.540414 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" podUID="cff0bb76-a914-4c38-90d2-f879b8c00b6e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.544042 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.544202 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.044179352 +0000 UTC m=+146.234707978 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.544376 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wbdv\" (UniqueName: \"kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.544433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.544512 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.544619 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.544875 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.044861243 +0000 UTC m=+146.235389959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.550189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" event={"ID":"3be12f5f-7fdb-4d42-b66e-42bccc2980e7","Type":"ContainerStarted","Data":"640689bbbdc1581889eceade0dab27f73328c44183c8780f04eb95905db5bda5"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.560022 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" event={"ID":"4c6e1d5c-4719-47d4-9faf-06d94241dca7","Type":"ContainerStarted","Data":"3e83e557caf471d9e8757bf4eee998d40693dc8556ca7d71516f6de414d1cde2"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.576929 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" event={"ID":"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63","Type":"ContainerStarted","Data":"472e424b61177722f952fad18f1e4a320fbaf89d90cd57a24435d51c476dfca1"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.576977 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" event={"ID":"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63","Type":"ContainerStarted","Data":"831c256bdd00490ca7b29e312135f2c8ae06bbdb02195fe7bb600d3a4fc0d49d"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.607255 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-7f488" podStartSLOduration=126.607238703 podStartE2EDuration="2m6.607238703s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.606744958 +0000 UTC m=+145.797273584" watchObservedRunningTime="2025-10-09 08:19:47.607238703 +0000 UTC m=+145.797767329" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.607409 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-5zq9q" podStartSLOduration=125.607405369 podStartE2EDuration="2m5.607405369s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.585317855 +0000 UTC m=+145.775846491" watchObservedRunningTime="2025-10-09 08:19:47.607405369 +0000 UTC m=+145.797933995" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.608730 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" event={"ID":"e7af426f-a554-434f-9735-922bebb651bf","Type":"ContainerStarted","Data":"64a519fc58892af16e77fdbb128ccb675437e613f54494630c9b0f8cb3a4136a"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.632194 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" event={"ID":"b03f0070-47ed-455e-b4b8-7a86fa5818af","Type":"ContainerStarted","Data":"eb5dd41a113c1b3357182b82302eec5661dc52959845534b7b565ae194b27be8"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.637357 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-dt452" podStartSLOduration=125.637343741 podStartE2EDuration="2m5.637343741s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.636105023 +0000 UTC m=+145.826633669" watchObservedRunningTime="2025-10-09 08:19:47.637343741 +0000 UTC m=+145.827872367" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.648226 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.648500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.648571 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wbdv\" (UniqueName: \"kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.648619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.649741 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.149725768 +0000 UTC m=+146.340254394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.650397 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.650886 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.657007 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" event={"ID":"edf09884-5c77-416e-b4ed-570657f567b6","Type":"ContainerStarted","Data":"a60a06e23781161036bdd26c6b81f404c6b7f363d520cde44f19ce02b14b4f82"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.657056 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" event={"ID":"edf09884-5c77-416e-b4ed-570657f567b6","Type":"ContainerStarted","Data":"6b2b73ae04906bffa39742fca57c6dac4792b64e9fb04bf956f3c5e11203423d"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.683756 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" podStartSLOduration=126.683741115 podStartE2EDuration="2m6.683741115s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.682005672 +0000 UTC m=+145.872534298" watchObservedRunningTime="2025-10-09 08:19:47.683741115 +0000 UTC m=+145.874269751" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.685672 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-cs9zz" event={"ID":"11d63870-f531-4733-a134-6de125f2a7a5","Type":"ContainerStarted","Data":"1ae69f07aa377fc2e4fcd29a1eed8db4592e5e7c659fa55f87b0c786889c8a01"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.689141 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.691025 4872 generic.go:334] "Generic (PLEG): container finished" podID="6f555e5b-289d-40e7-9671-c47c69282416" containerID="531604a396273217e72c593045071d4ab445508230a711e6f6bc9d1a2a541634" exitCode=0 Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.695461 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wbdv\" (UniqueName: \"kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv\") pod \"certified-operators-w6xb5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.707257 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" event={"ID":"6f555e5b-289d-40e7-9671-c47c69282416","Type":"ContainerDied","Data":"531604a396273217e72c593045071d4ab445508230a711e6f6bc9d1a2a541634"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.707313 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" event={"ID":"6f555e5b-289d-40e7-9671-c47c69282416","Type":"ContainerStarted","Data":"eb03d9f8f2d57fa7ed73266eb659f968f2ea0e921f02f9a47bf89645bae5a0b2"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.707325 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" event={"ID":"83fcb79e-6159-4129-8260-5314af9a6e8e","Type":"ContainerStarted","Data":"f7ca5fc6d4b278fa64da914872a0a0c0de5026ca301110937434c801ac5b58d0"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.707383 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.713477 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.721921 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" event={"ID":"50f906df-d560-4cd6-9e57-0f92a754c74a","Type":"ContainerStarted","Data":"12f0d004f6d2e79417950a42d2070d7b0733ce0cec0b02b387b4009b04fd7341"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.721965 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" event={"ID":"50f906df-d560-4cd6-9e57-0f92a754c74a","Type":"ContainerStarted","Data":"5b965ae718e8631d6016092201946fdbc4bba44fa4f0c99a9b34c8110cbc5116"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.722513 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.741427 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4gkkd" event={"ID":"2b9767b1-3531-4558-a6d7-55cae92ad568","Type":"ContainerStarted","Data":"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.741480 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4gkkd" event={"ID":"2b9767b1-3531-4558-a6d7-55cae92ad568","Type":"ContainerStarted","Data":"83d134f548d7652d63c0950088be313e24bc3c1d7c00e7efeeee79728b673d0b"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.758131 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dk2r4" podStartSLOduration=126.758115071 podStartE2EDuration="2m6.758115071s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.734269595 +0000 UTC m=+145.924798241" watchObservedRunningTime="2025-10-09 08:19:47.758115071 +0000 UTC m=+145.948643697" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.759672 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.763977 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.764031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.764216 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.764319 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh7bc\" (UniqueName: \"kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.766092 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.266080104 +0000 UTC m=+146.456608730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.782458 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-9dqkl" event={"ID":"e729bc27-1f38-4b33-8c06-ad200e076cb9","Type":"ContainerStarted","Data":"93ff5c4458d1b5183ce62b8680e8e1d75cae15e261a12765c27dc0ad454f1033"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.821126 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" podStartSLOduration=125.821105531 podStartE2EDuration="2m5.821105531s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.80697513 +0000 UTC m=+145.997503816" watchObservedRunningTime="2025-10-09 08:19:47.821105531 +0000 UTC m=+146.011634147" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.829169 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.844818 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" event={"ID":"dce7a5e8-3074-463c-9ff5-b68ad42ced94","Type":"ContainerStarted","Data":"2a42ace452389c7940e5cca5626968edfaebb0a6f02f6720ccf1fa076c560015"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.844875 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" event={"ID":"dce7a5e8-3074-463c-9ff5-b68ad42ced94","Type":"ContainerStarted","Data":"acc0304940e91a358d80c38e716fe02490408e03753258742f27f491e236c1ef"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.865563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.865795 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.865929 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.865969 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh7bc\" (UniqueName: \"kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.866305 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.366290208 +0000 UTC m=+146.556818834 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.866667 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.867992 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.889091 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" event={"ID":"81d1c103-d6b5-4e05-a262-10a658db5948","Type":"ContainerStarted","Data":"ff28cc9d93fc146e4e7954d976ef39367cead1000631a5cb4c361ab5120cea03"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.908657 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh7bc\" (UniqueName: \"kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc\") pod \"community-operators-ptfxz\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.910701 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.912152 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.915949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" event={"ID":"1f314406-7129-4061-922a-7c2fba4019d9","Type":"ContainerStarted","Data":"365e42c72f357ce35c2c2569c9f3ff47ea657d4e077e124e843dab6ebf44419c"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.933862 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-9dqkl" podStartSLOduration=7.933837456 podStartE2EDuration="7.933837456s" podCreationTimestamp="2025-10-09 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.928976208 +0000 UTC m=+146.119504834" watchObservedRunningTime="2025-10-09 08:19:47.933837456 +0000 UTC m=+146.124366082" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.933984 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.960198 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" event={"ID":"b973560f-edd3-4939-9f0e-146834366a2b","Type":"ContainerStarted","Data":"b6cb766ece63592a9083461ba04c76cd5ad2cab3fabebe1f4ac1b63ed9840c52"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.960262 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" event={"ID":"b973560f-edd3-4939-9f0e-146834366a2b","Type":"ContainerStarted","Data":"ae33194cef842c7137859a61e27c3af27eb09ca0060d1883fa8ab23d80a1523b"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.962441 4872 generic.go:334] "Generic (PLEG): container finished" podID="fd040384-e861-44c9-bffc-1dc10acb8e89" containerID="30a298986ecb7e022a1f54e5a48ce269d17616fd51ba3c5a9bb9ea8b6d7cf1bd" exitCode=0 Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.962487 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" event={"ID":"fd040384-e861-44c9-bffc-1dc10acb8e89","Type":"ContainerDied","Data":"30a298986ecb7e022a1f54e5a48ce269d17616fd51ba3c5a9bb9ea8b6d7cf1bd"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.962505 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" event={"ID":"fd040384-e861-44c9-bffc-1dc10acb8e89","Type":"ContainerStarted","Data":"ae655ba819c7620016e6f81dfa8744e095f4d97f54207bea9eb1294ba395d173"} Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.968803 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5k27\" (UniqueName: \"kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.968856 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.968923 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.969060 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:47 crc kubenswrapper[4872]: E1009 08:19:47.969382 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.469367919 +0000 UTC m=+146.659896545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:47 crc kubenswrapper[4872]: I1009 08:19:47.971756 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-fgzt2" podStartSLOduration=125.971739061 podStartE2EDuration="2m5.971739061s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:47.967919405 +0000 UTC m=+146.158448051" watchObservedRunningTime="2025-10-09 08:19:47.971739061 +0000 UTC m=+146.162267697" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.046881 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfq2v" event={"ID":"80dfc75a-1353-4b3a-9561-bb5532034038","Type":"ContainerStarted","Data":"4a120475ed43934144622a7e81d62b8ee6a5b0cfac7360f8f5a01e9ff3ef1ff2"} Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.047224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-dfq2v" event={"ID":"80dfc75a-1353-4b3a-9561-bb5532034038","Type":"ContainerStarted","Data":"e9fd4edaf8158dc1359f85072475f4bde922a4e71318b12bcabcdb6b098ccb7f"} Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.048012 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.048184 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.058467 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nm6xv" event={"ID":"192a1e23-c791-475f-be33-26f4b383a383","Type":"ContainerStarted","Data":"97991db77e87726ca428958ed934f451c29a2c57de08d08c50bcd15ff7589365"} Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.059271 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-4gkkd" podStartSLOduration=127.059249698 podStartE2EDuration="2m7.059249698s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:48.027232602 +0000 UTC m=+146.217761248" watchObservedRunningTime="2025-10-09 08:19:48.059249698 +0000 UTC m=+146.249778334" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.064443 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.065654 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.067376 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.067414 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.070082 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.070444 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5k27\" (UniqueName: \"kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.070468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.070559 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.071491 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.57147607 +0000 UTC m=+146.762004696 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.074336 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.074672 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.079110 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-g2q4g" event={"ID":"bb6c618e-f78b-40eb-8556-3824be52910d","Type":"ContainerStarted","Data":"55f1aed050d61c3613e692281802121975d42a8819ef81ed234f3e68c02e0b48"} Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.093936 4872 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-xlstb container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.093983 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.12:8080/healthz\": dial tcp 10.217.0.12:8080: connect: connection refused" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.102150 4872 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-5xj5z container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.102190 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" podUID="8a5e4235-b5ba-495b-8462-8796c968d1f4" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.17:8443/healthz\": dial tcp 10.217.0.17:8443: connect: connection refused" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.141754 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-bvcfj" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.142700 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-b5qvz" podStartSLOduration=127.142632469 podStartE2EDuration="2m7.142632469s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:48.140603487 +0000 UTC m=+146.331132113" watchObservedRunningTime="2025-10-09 08:19:48.142632469 +0000 UTC m=+146.333161095" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.147120 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-mqrzg" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.151457 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.178269 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbjsw\" (UniqueName: \"kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.178505 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.178569 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.178631 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.187918 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.687904148 +0000 UTC m=+146.878432764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.227808 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:48 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:48 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:48 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.227851 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.244438 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5k27\" (UniqueName: \"kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27\") pod \"certified-operators-lvbdw\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.280247 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.280522 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbjsw\" (UniqueName: \"kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.280582 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.280605 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.281202 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.281590 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.781562363 +0000 UTC m=+146.972090989 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.281860 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.293912 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.349825 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbjsw\" (UniqueName: \"kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw\") pod \"community-operators-dhnbs\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.382802 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.383477 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.883465008 +0000 UTC m=+147.073993624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.399322 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-6k8nz" podStartSLOduration=127.399303031 podStartE2EDuration="2m7.399303031s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:48.292278509 +0000 UTC m=+146.482807135" watchObservedRunningTime="2025-10-09 08:19:48.399303031 +0000 UTC m=+146.589831657" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.431802 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h2ll5" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.441903 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.493286 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.493909 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:48.993881633 +0000 UTC m=+147.184410259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.527188 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-bkhr6" podStartSLOduration=127.527168897 podStartE2EDuration="2m7.527168897s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:48.425820509 +0000 UTC m=+146.616349155" watchObservedRunningTime="2025-10-09 08:19:48.527168897 +0000 UTC m=+146.717697523" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.590780 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.595389 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.596670 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.096655705 +0000 UTC m=+147.287184331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.699437 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-dfq2v" podStartSLOduration=8.699422516 podStartE2EDuration="8.699422516s" podCreationTimestamp="2025-10-09 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:48.672417954 +0000 UTC m=+146.862946580" watchObservedRunningTime="2025-10-09 08:19:48.699422516 +0000 UTC m=+146.889951142" Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.715344 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.715761 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.215747034 +0000 UTC m=+147.406275660 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.808215 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.817502 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.817795 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.317783453 +0000 UTC m=+147.508312079 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.919065 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:48 crc kubenswrapper[4872]: E1009 08:19:48.919679 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.419659998 +0000 UTC m=+147.610188624 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:48 crc kubenswrapper[4872]: I1009 08:19:48.988067 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.024470 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.024895 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.524881724 +0000 UTC m=+147.715410350 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.083156 4872 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-g747n container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.083695 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.106335 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" event={"ID":"a1a5b7f5-7caa-4fe2-afb4-6c7176939e63","Type":"ContainerStarted","Data":"de21db3df007b83cf9a0997350b7f8ed9a3d12aa811fd3337b17fbb06c5f9f09"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.119836 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerStarted","Data":"a8e47bce071f2cf5bdf5cc4167d283ef71681300b1bce37a68f0034c911699fc"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.125028 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.125468 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.625453579 +0000 UTC m=+147.815982205 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.141937 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-wcb54" podStartSLOduration=127.141917941 podStartE2EDuration="2m7.141917941s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:49.139992232 +0000 UTC m=+147.330520858" watchObservedRunningTime="2025-10-09 08:19:49.141917941 +0000 UTC m=+147.332446567" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.156247 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" event={"ID":"fd040384-e861-44c9-bffc-1dc10acb8e89","Type":"ContainerStarted","Data":"e96888fc532cd1d328dc733f3f0d160f2cd97804f9d6e25145c900318fb42181"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.171672 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerStarted","Data":"4454fc1f0fb1d18e60c99630f8ca6e3d35bfdaca2a48a7de07458ddd3f1b5d6d"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.171715 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerStarted","Data":"852d683e3d62168542238cc114530f0e5f66b95f319e35679484e779a890c954"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.175988 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.183701 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" podStartSLOduration=127.183687704 podStartE2EDuration="2m7.183687704s" podCreationTimestamp="2025-10-09 08:17:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:49.1816036 +0000 UTC m=+147.372132226" watchObservedRunningTime="2025-10-09 08:19:49.183687704 +0000 UTC m=+147.374216330" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.190263 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" event={"ID":"6f555e5b-289d-40e7-9671-c47c69282416","Type":"ContainerStarted","Data":"8aa2faea5384cb4739dc4875851b3249693814962b8ddb24ee494472aa24fe22"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.193207 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerStarted","Data":"aa8860dd9434f67289f5161c31b0a3a6c3d28aa2bd3ca6d056d00f6763e266c4"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.195058 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" event={"ID":"83fcb79e-6159-4129-8260-5314af9a6e8e","Type":"ContainerStarted","Data":"7ef6cccff1a860c704dd04562f7ee8606e41cd3bedff872fcfc563d5afd8ead1"} Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.199039 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.199096 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.222818 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:49 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:49 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:49 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.222870 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.228839 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.232023 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.732010086 +0000 UTC m=+147.922538712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.334362 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.334985 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.335250 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.335332 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.335412 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.335872 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.336698 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.836683896 +0000 UTC m=+148.027212522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.339259 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.350350 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.350932 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.354462 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:49 crc kubenswrapper[4872]: W1009 08:19:49.358791 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71f030ad_a0a1_49b8_9910_1bf39006e56a.slice/crio-b9707481ebb6894284443d9e5f862c48cc89969c6f02e19026e9465f4d68c43c WatchSource:0}: Error finding container b9707481ebb6894284443d9e5f862c48cc89969c6f02e19026e9465f4d68c43c: Status 404 returned error can't find the container with id b9707481ebb6894284443d9e5f862c48cc89969c6f02e19026e9465f4d68c43c Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.384841 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.393040 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.436950 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.437292 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:49.937279432 +0000 UTC m=+148.127808058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.462690 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-5xj5z" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.484527 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.521865 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.521916 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.526888 4872 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-7cglv container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.526984 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" podUID="fd040384-e861-44c9-bffc-1dc10acb8e89" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.8:8443/livez\": dial tcp 10.217.0.8:8443: connect: connection refused" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.535761 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.538992 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.540181 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.541468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.552020 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.051994947 +0000 UTC m=+148.242523573 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.553001 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.553182 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.652305 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.653861 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.655336 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.655389 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.655459 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.655819 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.155805321 +0000 UTC m=+148.346333947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.666913 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.667151 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756062 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.756241 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.256222421 +0000 UTC m=+148.446751047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756377 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756437 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756472 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756535 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756553 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bcbt\" (UniqueName: \"kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.756997 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.757237 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.257229822 +0000 UTC m=+148.447758448 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.779425 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.857975 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.858114 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.858172 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.858190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bcbt\" (UniqueName: \"kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.858535 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.358519228 +0000 UTC m=+148.549047854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.858911 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.859078 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.880311 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bcbt\" (UniqueName: \"kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt\") pod \"redhat-marketplace-jds9b\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.906475 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:49 crc kubenswrapper[4872]: W1009 08:19:49.956816 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-f09aae868eae3a07e5bbbd78aa227bc50de731d13505b07a7a6e5b5b8e907f06 WatchSource:0}: Error finding container f09aae868eae3a07e5bbbd78aa227bc50de731d13505b07a7a6e5b5b8e907f06: Status 404 returned error can't find the container with id f09aae868eae3a07e5bbbd78aa227bc50de731d13505b07a7a6e5b5b8e907f06 Oct 09 08:19:49 crc kubenswrapper[4872]: I1009 08:19:49.960608 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:49 crc kubenswrapper[4872]: E1009 08:19:49.960955 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.46094304 +0000 UTC m=+148.651471666 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.044772 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.045820 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.048311 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.070662 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.072083 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.071429 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.571400726 +0000 UTC m=+148.761929352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.173403 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.173828 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.673808367 +0000 UTC m=+148.864336993 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.173893 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkzr8\" (UniqueName: \"kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.174028 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.174087 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.215900 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:50 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:50 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:50 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.215952 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.251183 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"f09aae868eae3a07e5bbbd78aa227bc50de731d13505b07a7a6e5b5b8e907f06"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.268675 4872 generic.go:334] "Generic (PLEG): container finished" podID="c4f8212a-5309-4088-b643-4f3d421e8799" containerID="dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61" exitCode=0 Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.268787 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerDied","Data":"dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.277966 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.278145 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.278172 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.278227 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkzr8\" (UniqueName: \"kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.278462 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.778448195 +0000 UTC m=+148.968976821 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.279839 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.280044 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.310577 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkzr8\" (UniqueName: \"kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8\") pod \"redhat-marketplace-blvp8\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.335417 4872 generic.go:334] "Generic (PLEG): container finished" podID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerID="1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061" exitCode=0 Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.335477 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerDied","Data":"1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.335500 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerStarted","Data":"b9707481ebb6894284443d9e5f862c48cc89969c6f02e19026e9465f4d68c43c"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.337465 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.358416 4872 generic.go:334] "Generic (PLEG): container finished" podID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerID="4454fc1f0fb1d18e60c99630f8ca6e3d35bfdaca2a48a7de07458ddd3f1b5d6d" exitCode=0 Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.358507 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerDied","Data":"4454fc1f0fb1d18e60c99630f8ca6e3d35bfdaca2a48a7de07458ddd3f1b5d6d"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.379570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.380315 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.880300789 +0000 UTC m=+149.070829425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.390239 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.402427 4872 generic.go:334] "Generic (PLEG): container finished" podID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerID="61ef1e26f3519f47ed3ca5c70acc01f7d82f27e3599b4abf3dbad4d470d1d4d4" exitCode=0 Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.402478 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerDied","Data":"61ef1e26f3519f47ed3ca5c70acc01f7d82f27e3599b4abf3dbad4d470d1d4d4"} Oct 09 08:19:50 crc kubenswrapper[4872]: W1009 08:19:50.402549 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-827d2f6e6eb21c11603a5c9ca0fe9f1d9dea37b9a9bf963de233a9c7f286f40c WatchSource:0}: Error finding container 827d2f6e6eb21c11603a5c9ca0fe9f1d9dea37b9a9bf963de233a9c7f286f40c: Status 404 returned error can't find the container with id 827d2f6e6eb21c11603a5c9ca0fe9f1d9dea37b9a9bf963de233a9c7f286f40c Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.404385 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"abfe9a58ea3822baf0754ab8b7ea035f8ed1f3155262266ac0aa7972020b7f94"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.409116 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" event={"ID":"6f555e5b-289d-40e7-9671-c47c69282416","Type":"ContainerStarted","Data":"54afd5833633b59ab9d6750286ad4a9d4f7d1fdf00e03f1ad762392d4bdd7d6b"} Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.481103 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.482565 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:50.982548715 +0000 UTC m=+149.173077341 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.582996 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.584992 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.084974276 +0000 UTC m=+149.275502902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.641691 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" podStartSLOduration=129.641676174 podStartE2EDuration="2m9.641676174s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:50.435491661 +0000 UTC m=+148.626020297" watchObservedRunningTime="2025-10-09 08:19:50.641676174 +0000 UTC m=+148.832204810" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.642990 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.644222 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.674988 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.685160 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.685326 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.185299044 +0000 UTC m=+149.375827670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.704580 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.732714 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.733153 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.233138341 +0000 UTC m=+149.423666967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.790240 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.835262 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.835445 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcfmq\" (UniqueName: \"kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.835479 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.835511 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.835676 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.335654695 +0000 UTC m=+149.526183321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.835820 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.836208 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.336199772 +0000 UTC m=+149.526728398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.898274 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.940173 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.940483 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcfmq\" (UniqueName: \"kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.940524 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.940562 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.941130 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: E1009 08:19:50.941221 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.441203182 +0000 UTC m=+149.631731808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.941795 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.980440 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcfmq\" (UniqueName: \"kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq\") pod \"redhat-operators-5w56c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:50 crc kubenswrapper[4872]: I1009 08:19:50.980765 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.041843 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.042723 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.542707485 +0000 UTC m=+149.733236111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.047502 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.050240 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.090944 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.144894 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.644866937 +0000 UTC m=+149.835395563 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.144966 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.145500 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.645491356 +0000 UTC m=+149.836019982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.145683 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.220021 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:51 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:51 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:51 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.220086 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.246730 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.246950 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5bg6\" (UniqueName: \"kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.246991 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.746969589 +0000 UTC m=+149.937498215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.247048 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.247155 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.247236 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.247299 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.747285098 +0000 UTC m=+149.937813724 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.341220 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.347934 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.348120 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.84809146 +0000 UTC m=+150.038620086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.348171 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.348290 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.348393 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5bg6\" (UniqueName: \"kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.348450 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.348817 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.848809202 +0000 UTC m=+150.039337828 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.349270 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.349320 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.371809 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5bg6\" (UniqueName: \"kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6\") pod \"redhat-operators-vz87x\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.394794 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.415457 4872 generic.go:334] "Generic (PLEG): container finished" podID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerID="4fa4bfd3fce9ca84fb76d1d7799db694e6e0a2cf57102874ebd71d3b90ebfe6e" exitCode=0 Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.415546 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerDied","Data":"4fa4bfd3fce9ca84fb76d1d7799db694e6e0a2cf57102874ebd71d3b90ebfe6e"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.415667 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerStarted","Data":"b41545544afdccfb640576a1cf20ad2cba40486e076a74af31055c29ef440a24"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.422962 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"69de4cded4ac7f5893030c835e504df56d334c9c270192fa350c99ffcda79677"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.423010 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"827d2f6e6eb21c11603a5c9ca0fe9f1d9dea37b9a9bf963de233a9c7f286f40c"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.423246 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.426893 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc677e74-a252-411e-8ca3-ca581fab03b6","Type":"ContainerStarted","Data":"c517ffa170342ec77c0d9d611fc276c8fae077919411b3cd477c1b1f6239d265"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.426940 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc677e74-a252-411e-8ca3-ca581fab03b6","Type":"ContainerStarted","Data":"42507af467f49904286c85b5a5ad78f34944d756f41dc7c01495485d3afb2672"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.428277 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerStarted","Data":"0931e919ea5a7af76bfe120bbe5a3c75fe57e69bb55e3285f6f133a4e81b0350"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.429431 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerStarted","Data":"1f93d2325f3feb05af21cfffc633091e0874adc9715753e4b62dbe4404f21957"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.437854 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"bbf26b8d246557943cbb24b46baf2c5b39c01b25cc48d3c7f9d696b538c590a0"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.440416 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8a5fefc7f1d403e6922238a8a754bd340237582ebaff38ce2cc4597eba57c93b"} Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.444234 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.44421775 podStartE2EDuration="2.44421775s" podCreationTimestamp="2025-10-09 08:19:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:51.443664903 +0000 UTC m=+149.634193529" watchObservedRunningTime="2025-10-09 08:19:51.44421775 +0000 UTC m=+149.634746376" Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.449391 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.449512 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.94949731 +0000 UTC m=+150.140025936 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.449690 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.449997 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:51.949989246 +0000 UTC m=+150.140517872 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.552554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.552835 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.052807399 +0000 UTC m=+150.243336025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.552971 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.555135 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.055117829 +0000 UTC m=+150.245646515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.657374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.658203 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.15818337 +0000 UTC m=+150.348711996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.757930 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.759664 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.760038 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.260022273 +0000 UTC m=+150.450550899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: W1009 08:19:51.765303 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e121152_fd8c_42a6_98a5_d7ea05dfab3e.slice/crio-0f88be4b07c82b5ca709b3636d817b9938f9cdf39160bc2273730f9d695360c1 WatchSource:0}: Error finding container 0f88be4b07c82b5ca709b3636d817b9938f9cdf39160bc2273730f9d695360c1: Status 404 returned error can't find the container with id 0f88be4b07c82b5ca709b3636d817b9938f9cdf39160bc2273730f9d695360c1 Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.861090 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.861239 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.361207027 +0000 UTC m=+150.551735653 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.861369 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.861737 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.361726933 +0000 UTC m=+150.552255559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.962469 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.962667 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.462623167 +0000 UTC m=+150.653151803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:51 crc kubenswrapper[4872]: I1009 08:19:51.963118 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:51 crc kubenswrapper[4872]: E1009 08:19:51.963572 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.463544165 +0000 UTC m=+150.654072791 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.051989 4872 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.064113 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.064256 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.564237944 +0000 UTC m=+150.754766570 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.064347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.064720 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.564712728 +0000 UTC m=+150.755241354 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.165883 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.166291 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.666260323 +0000 UTC m=+150.856788949 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.166728 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.167113 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.667096588 +0000 UTC m=+150.857625214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.219597 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:52 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:52 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:52 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.219680 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.268393 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.268551 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.768526669 +0000 UTC m=+150.959055295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.268763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.269105 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.769090607 +0000 UTC m=+150.959619233 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.370398 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.370861 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.870831867 +0000 UTC m=+151.061360493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.491422 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.492197 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:52.992180595 +0000 UTC m=+151.182709221 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.494853 4872 generic.go:334] "Generic (PLEG): container finished" podID="bc677e74-a252-411e-8ca3-ca581fab03b6" containerID="c517ffa170342ec77c0d9d611fc276c8fae077919411b3cd477c1b1f6239d265" exitCode=0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.507119 4872 generic.go:334] "Generic (PLEG): container finished" podID="b03f0070-47ed-455e-b4b8-7a86fa5818af" containerID="eb5dd41a113c1b3357182b82302eec5661dc52959845534b7b565ae194b27be8" exitCode=0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.511040 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc677e74-a252-411e-8ca3-ca581fab03b6","Type":"ContainerDied","Data":"c517ffa170342ec77c0d9d611fc276c8fae077919411b3cd477c1b1f6239d265"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.511072 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" event={"ID":"b03f0070-47ed-455e-b4b8-7a86fa5818af","Type":"ContainerDied","Data":"eb5dd41a113c1b3357182b82302eec5661dc52959845534b7b565ae194b27be8"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.511400 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerID="922e79b9edbb83e33796b0dd57f6f8aaba5d2f182a9c3164cb8a0d3ffe6174e9" exitCode=0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.511443 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerDied","Data":"922e79b9edbb83e33796b0dd57f6f8aaba5d2f182a9c3164cb8a0d3ffe6174e9"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.514385 4872 generic.go:334] "Generic (PLEG): container finished" podID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerID="e15be488cd16c8db5e924abb709f534403245b1a7366bbedac3cb83f210d7d56" exitCode=0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.514421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerDied","Data":"e15be488cd16c8db5e924abb709f534403245b1a7366bbedac3cb83f210d7d56"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.526761 4872 generic.go:334] "Generic (PLEG): container finished" podID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerID="4fb50b9d16a3ce8520609c47d06d6c51132368670bbba8aac8a1cba97cdfa7f9" exitCode=0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.526880 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerDied","Data":"4fb50b9d16a3ce8520609c47d06d6c51132368670bbba8aac8a1cba97cdfa7f9"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.526934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerStarted","Data":"0f88be4b07c82b5ca709b3636d817b9938f9cdf39160bc2273730f9d695360c1"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.552477 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" event={"ID":"83fcb79e-6159-4129-8260-5314af9a6e8e","Type":"ContainerStarted","Data":"21542121afceb10592ce0188f7e00727592ab6203daedf269215b23259c3a997"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.552521 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" event={"ID":"83fcb79e-6159-4129-8260-5314af9a6e8e","Type":"ContainerStarted","Data":"4ce1387ab861d8dc53092f06c58861bb837035cbbf98de0c43ef21f85fcbbb78"} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.592211 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.592398 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-09 08:19:53.092366588 +0000 UTC m=+151.282895214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.592509 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: E1009 08:19:52.592998 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-09 08:19:53.092876004 +0000 UTC m=+151.283404630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-78fqw" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.657269 4872 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-09T08:19:52.052028692Z","Handler":null,"Name":""} Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.675831 4872 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.675876 4872 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.695205 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.699942 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.797095 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.802194 4872 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.802240 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:52 crc kubenswrapper[4872]: I1009 08:19:52.896795 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-78fqw\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.122220 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.177064 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.182264 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.225013 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.226486 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:53 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:53 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:53 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.226539 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.528230 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.575051 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" event={"ID":"83fcb79e-6159-4129-8260-5314af9a6e8e","Type":"ContainerStarted","Data":"2dd3c9f9154d86e34ef4b1f916224965bdbf876f7d293bf0de8038ce094f990b"} Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.595182 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-fbd5f" podStartSLOduration=13.595161057 podStartE2EDuration="13.595161057s" podCreationTimestamp="2025-10-09 08:19:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:53.590750392 +0000 UTC m=+151.781279028" watchObservedRunningTime="2025-10-09 08:19:53.595161057 +0000 UTC m=+151.785689693" Oct 09 08:19:53 crc kubenswrapper[4872]: W1009 08:19:53.659264 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd734d590_eccc_42cc_ad8d_14e28f3c1882.slice/crio-c9f95989b6ade0b2ea2872e9f33e98eeaf9affb8c00bb0aaecc79925ccd09881 WatchSource:0}: Error finding container c9f95989b6ade0b2ea2872e9f33e98eeaf9affb8c00bb0aaecc79925ccd09881: Status 404 returned error can't find the container with id c9f95989b6ade0b2ea2872e9f33e98eeaf9affb8c00bb0aaecc79925ccd09881 Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.940797 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:53 crc kubenswrapper[4872]: I1009 08:19:53.941568 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.060234 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 08:19:54 crc kubenswrapper[4872]: E1009 08:19:54.060487 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b03f0070-47ed-455e-b4b8-7a86fa5818af" containerName="collect-profiles" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.060502 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b03f0070-47ed-455e-b4b8-7a86fa5818af" containerName="collect-profiles" Oct 09 08:19:54 crc kubenswrapper[4872]: E1009 08:19:54.060515 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc677e74-a252-411e-8ca3-ca581fab03b6" containerName="pruner" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.060523 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc677e74-a252-411e-8ca3-ca581fab03b6" containerName="pruner" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.060893 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b03f0070-47ed-455e-b4b8-7a86fa5818af" containerName="collect-profiles" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.060915 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc677e74-a252-411e-8ca3-ca581fab03b6" containerName="pruner" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.061336 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.063375 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.065897 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.072521 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.115593 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume\") pod \"b03f0070-47ed-455e-b4b8-7a86fa5818af\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.115703 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume\") pod \"b03f0070-47ed-455e-b4b8-7a86fa5818af\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.115742 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-258cc\" (UniqueName: \"kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc\") pod \"b03f0070-47ed-455e-b4b8-7a86fa5818af\" (UID: \"b03f0070-47ed-455e-b4b8-7a86fa5818af\") " Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.115790 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir\") pod \"bc677e74-a252-411e-8ca3-ca581fab03b6\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.115837 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access\") pod \"bc677e74-a252-411e-8ca3-ca581fab03b6\" (UID: \"bc677e74-a252-411e-8ca3-ca581fab03b6\") " Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.117270 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume" (OuterVolumeSpecName: "config-volume") pod "b03f0070-47ed-455e-b4b8-7a86fa5818af" (UID: "b03f0070-47ed-455e-b4b8-7a86fa5818af"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.117356 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bc677e74-a252-411e-8ca3-ca581fab03b6" (UID: "bc677e74-a252-411e-8ca3-ca581fab03b6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.126665 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc" (OuterVolumeSpecName: "kube-api-access-258cc") pod "b03f0070-47ed-455e-b4b8-7a86fa5818af" (UID: "b03f0070-47ed-455e-b4b8-7a86fa5818af"). InnerVolumeSpecName "kube-api-access-258cc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.127627 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bc677e74-a252-411e-8ca3-ca581fab03b6" (UID: "bc677e74-a252-411e-8ca3-ca581fab03b6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.129283 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b03f0070-47ed-455e-b4b8-7a86fa5818af" (UID: "b03f0070-47ed-455e-b4b8-7a86fa5818af"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.217609 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.217686 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.218118 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bc677e74-a252-411e-8ca3-ca581fab03b6-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.218134 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b03f0070-47ed-455e-b4b8-7a86fa5818af-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.218143 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b03f0070-47ed-455e-b4b8-7a86fa5818af-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.218151 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-258cc\" (UniqueName: \"kubernetes.io/projected/b03f0070-47ed-455e-b4b8-7a86fa5818af-kube-api-access-258cc\") on node \"crc\" DevicePath \"\"" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.218161 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bc677e74-a252-411e-8ca3-ca581fab03b6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.235916 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:54 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:54 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:54 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.235987 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.284851 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.284926 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.284903 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.284993 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.289315 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.289415 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.301705 4872 patch_prober.go:28] interesting pod/console-f9d7485db-4gkkd container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.301768 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-4gkkd" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerName="console" probeResult="failure" output="Get \"https://10.217.0.28:8443/health\": dial tcp 10.217.0.28:8443: connect: connection refused" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.319118 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.319214 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.319221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.348998 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.403070 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.454609 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.455031 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.503753 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.504655 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.530476 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.539575 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7cglv" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.651625 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"bc677e74-a252-411e-8ca3-ca581fab03b6","Type":"ContainerDied","Data":"42507af467f49904286c85b5a5ad78f34944d756f41dc7c01495485d3afb2672"} Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.651685 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="42507af467f49904286c85b5a5ad78f34944d756f41dc7c01495485d3afb2672" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.651770 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.666235 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.666241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl" event={"ID":"b03f0070-47ed-455e-b4b8-7a86fa5818af","Type":"ContainerDied","Data":"22dceb6618e8d91e18edbf7f8f0f104a78fc4d01b2ceb1b5be2bf7ada304cc49"} Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.666290 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="22dceb6618e8d91e18edbf7f8f0f104a78fc4d01b2ceb1b5be2bf7ada304cc49" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.673169 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" event={"ID":"d734d590-eccc-42cc-ad8d-14e28f3c1882","Type":"ContainerStarted","Data":"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60"} Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.673217 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" event={"ID":"d734d590-eccc-42cc-ad8d-14e28f3c1882","Type":"ContainerStarted","Data":"c9f95989b6ade0b2ea2872e9f33e98eeaf9affb8c00bb0aaecc79925ccd09881"} Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.680424 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-rlfqc" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.711959 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" podStartSLOduration=133.711938778 podStartE2EDuration="2m13.711938778s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:19:54.693697502 +0000 UTC m=+152.884226128" watchObservedRunningTime="2025-10-09 08:19:54.711938778 +0000 UTC m=+152.902467414" Oct 09 08:19:54 crc kubenswrapper[4872]: I1009 08:19:54.980845 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 09 08:19:55 crc kubenswrapper[4872]: I1009 08:19:55.220824 4872 patch_prober.go:28] interesting pod/router-default-5444994796-gsqtq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 09 08:19:55 crc kubenswrapper[4872]: [-]has-synced failed: reason withheld Oct 09 08:19:55 crc kubenswrapper[4872]: [+]process-running ok Oct 09 08:19:55 crc kubenswrapper[4872]: healthz check failed Oct 09 08:19:55 crc kubenswrapper[4872]: I1009 08:19:55.220896 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gsqtq" podUID="61adb101-0d3a-4018-aa4f-4ed4a435148b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 09 08:19:55 crc kubenswrapper[4872]: I1009 08:19:55.724102 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"54fadc33-526c-4803-a659-90cac0c4a861","Type":"ContainerStarted","Data":"11612cbc82322a8f1f8670eb4f81988bf04252792495134c072a6e61c20075f7"} Oct 09 08:19:55 crc kubenswrapper[4872]: I1009 08:19:55.724448 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:19:56 crc kubenswrapper[4872]: I1009 08:19:56.217759 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:56 crc kubenswrapper[4872]: I1009 08:19:56.223282 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gsqtq" Oct 09 08:19:56 crc kubenswrapper[4872]: I1009 08:19:56.331439 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-dfq2v" Oct 09 08:19:56 crc kubenswrapper[4872]: I1009 08:19:56.738689 4872 generic.go:334] "Generic (PLEG): container finished" podID="54fadc33-526c-4803-a659-90cac0c4a861" containerID="479b0d3c63f3f64adbc7be0b13e060d0d63d137c6b1eb83b5637000e3983d24a" exitCode=0 Oct 09 08:19:56 crc kubenswrapper[4872]: I1009 08:19:56.739674 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"54fadc33-526c-4803-a659-90cac0c4a861","Type":"ContainerDied","Data":"479b0d3c63f3f64adbc7be0b13e060d0d63d137c6b1eb83b5637000e3983d24a"} Oct 09 08:20:01 crc kubenswrapper[4872]: I1009 08:20:01.040359 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:20:03 crc kubenswrapper[4872]: I1009 08:20:03.711235 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:20:03 crc kubenswrapper[4872]: I1009 08:20:03.724555 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/326d4497-ac5f-4a64-afde-14ded51665b8-metrics-certs\") pod \"network-metrics-daemon-8grb6\" (UID: \"326d4497-ac5f-4a64-afde-14ded51665b8\") " pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:20:03 crc kubenswrapper[4872]: I1009 08:20:03.777602 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-8grb6" Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.281303 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.281704 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.281352 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.281822 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.294278 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.298784 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:20:04 crc kubenswrapper[4872]: I1009 08:20:04.909614 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.029899 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access\") pod \"54fadc33-526c-4803-a659-90cac0c4a861\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.030054 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir\") pod \"54fadc33-526c-4803-a659-90cac0c4a861\" (UID: \"54fadc33-526c-4803-a659-90cac0c4a861\") " Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.030190 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "54fadc33-526c-4803-a659-90cac0c4a861" (UID: "54fadc33-526c-4803-a659-90cac0c4a861"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.030357 4872 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/54fadc33-526c-4803-a659-90cac0c4a861-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.041388 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "54fadc33-526c-4803-a659-90cac0c4a861" (UID: "54fadc33-526c-4803-a659-90cac0c4a861"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.131669 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54fadc33-526c-4803-a659-90cac0c4a861-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.812735 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"54fadc33-526c-4803-a659-90cac0c4a861","Type":"ContainerDied","Data":"11612cbc82322a8f1f8670eb4f81988bf04252792495134c072a6e61c20075f7"} Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.813029 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11612cbc82322a8f1f8670eb4f81988bf04252792495134c072a6e61c20075f7" Oct 09 08:20:05 crc kubenswrapper[4872]: I1009 08:20:05.812804 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 09 08:20:08 crc kubenswrapper[4872]: I1009 08:20:08.157126 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:20:08 crc kubenswrapper[4872]: I1009 08:20:08.157187 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:20:13 crc kubenswrapper[4872]: I1009 08:20:13.185435 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.282504 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.283004 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.283064 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.283872 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"97991db77e87726ca428958ed934f451c29a2c57de08d08c50bcd15ff7589365"} pod="openshift-console/downloads-7954f5f757-nm6xv" containerMessage="Container download-server failed liveness probe, will be restarted" Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.284002 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" containerID="cri-o://97991db77e87726ca428958ed934f451c29a2c57de08d08c50bcd15ff7589365" gracePeriod=2 Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.284398 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.284503 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.285187 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:14 crc kubenswrapper[4872]: I1009 08:20:14.285232 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:16 crc kubenswrapper[4872]: I1009 08:20:16.450820 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-8grb6"] Oct 09 08:20:16 crc kubenswrapper[4872]: I1009 08:20:16.883738 4872 generic.go:334] "Generic (PLEG): container finished" podID="192a1e23-c791-475f-be33-26f4b383a383" containerID="97991db77e87726ca428958ed934f451c29a2c57de08d08c50bcd15ff7589365" exitCode=0 Oct 09 08:20:16 crc kubenswrapper[4872]: I1009 08:20:16.883786 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nm6xv" event={"ID":"192a1e23-c791-475f-be33-26f4b383a383","Type":"ContainerDied","Data":"97991db77e87726ca428958ed934f451c29a2c57de08d08c50bcd15ff7589365"} Oct 09 08:20:17 crc kubenswrapper[4872]: E1009 08:20:17.082314 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 09 08:20:17 crc kubenswrapper[4872]: E1009 08:20:17.082561 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4bcbt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-jds9b_openshift-marketplace(034eb0f7-581e-4e3d-aa86-12d7dc73690a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:17 crc kubenswrapper[4872]: E1009 08:20:17.083730 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-jds9b" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" Oct 09 08:20:24 crc kubenswrapper[4872]: I1009 08:20:24.170871 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-2775d" Oct 09 08:20:24 crc kubenswrapper[4872]: I1009 08:20:24.281322 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:24 crc kubenswrapper[4872]: I1009 08:20:24.281406 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:25 crc kubenswrapper[4872]: E1009 08:20:25.712008 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-jds9b" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" Oct 09 08:20:28 crc kubenswrapper[4872]: E1009 08:20:28.441091 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 09 08:20:28 crc kubenswrapper[4872]: E1009 08:20:28.441508 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-f5bg6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vz87x_openshift-marketplace(0e121152-fd8c-42a6-98a5-d7ea05dfab3e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:28 crc kubenswrapper[4872]: E1009 08:20:28.442576 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-vz87x" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" Oct 09 08:20:29 crc kubenswrapper[4872]: I1009 08:20:29.549266 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.165892 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.166347 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lcfmq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-5w56c_openshift-marketplace(fc1ccd35-725f-453b-988e-0d2da9b17b6c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.167563 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-5w56c" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.568315 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.568488 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5k27,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-lvbdw_openshift-marketplace(71f030ad-a0a1-49b8-9910-1bf39006e56a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.569583 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-lvbdw" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.649545 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.649700 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hkzr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-blvp8_openshift-marketplace(a5c8e529-78d6-4132-b94e-91df9aadda37): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:32 crc kubenswrapper[4872]: E1009 08:20:32.650913 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-blvp8" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" Oct 09 08:20:34 crc kubenswrapper[4872]: E1009 08:20:34.038319 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 09 08:20:34 crc kubenswrapper[4872]: E1009 08:20:34.041061 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wbdv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-w6xb5_openshift-marketplace(8839fa4b-4de5-4399-a83d-4c3cced15cd5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:34 crc kubenswrapper[4872]: E1009 08:20:34.042518 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-w6xb5" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" Oct 09 08:20:34 crc kubenswrapper[4872]: I1009 08:20:34.282509 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:34 crc kubenswrapper[4872]: I1009 08:20:34.282568 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:36 crc kubenswrapper[4872]: W1009 08:20:36.780874 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod326d4497_ac5f_4a64_afde_14ded51665b8.slice/crio-6deb91020fde935ca7d5b049e9df5ec74111151b3dc884394a5c737645509ffd WatchSource:0}: Error finding container 6deb91020fde935ca7d5b049e9df5ec74111151b3dc884394a5c737645509ffd: Status 404 returned error can't find the container with id 6deb91020fde935ca7d5b049e9df5ec74111151b3dc884394a5c737645509ffd Oct 09 08:20:36 crc kubenswrapper[4872]: E1009 08:20:36.781135 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-w6xb5" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" Oct 09 08:20:36 crc kubenswrapper[4872]: E1009 08:20:36.781214 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-5w56c" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" Oct 09 08:20:36 crc kubenswrapper[4872]: E1009 08:20:36.781215 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-blvp8" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" Oct 09 08:20:36 crc kubenswrapper[4872]: E1009 08:20:36.781308 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-lvbdw" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" Oct 09 08:20:36 crc kubenswrapper[4872]: E1009 08:20:36.781344 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vz87x" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" Oct 09 08:20:36 crc kubenswrapper[4872]: I1009 08:20:36.982235 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8grb6" event={"ID":"326d4497-ac5f-4a64-afde-14ded51665b8","Type":"ContainerStarted","Data":"6deb91020fde935ca7d5b049e9df5ec74111151b3dc884394a5c737645509ffd"} Oct 09 08:20:37 crc kubenswrapper[4872]: I1009 08:20:37.991541 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8grb6" event={"ID":"326d4497-ac5f-4a64-afde-14ded51665b8","Type":"ContainerStarted","Data":"d5274590ac0cd6792709ceb324eea004ad8eddb493094b3626adee7ed7a874dd"} Oct 09 08:20:37 crc kubenswrapper[4872]: I1009 08:20:37.995270 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-nm6xv" event={"ID":"192a1e23-c791-475f-be33-26f4b383a383","Type":"ContainerStarted","Data":"28ab9496a6e78a12036e5a2d7928fad583c5cbb510a3e465242bb328d6a0ada1"} Oct 09 08:20:38 crc kubenswrapper[4872]: E1009 08:20:38.150752 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 09 08:20:38 crc kubenswrapper[4872]: E1009 08:20:38.150940 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hh7bc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ptfxz_openshift-marketplace(933f1f0c-fc00-4ca8-b364-d97cebb40502): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:38 crc kubenswrapper[4872]: E1009 08:20:38.152436 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ptfxz" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" Oct 09 08:20:38 crc kubenswrapper[4872]: I1009 08:20:38.157321 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:20:38 crc kubenswrapper[4872]: I1009 08:20:38.157369 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:20:39 crc kubenswrapper[4872]: I1009 08:20:39.002204 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:20:39 crc kubenswrapper[4872]: I1009 08:20:39.002371 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:39 crc kubenswrapper[4872]: I1009 08:20:39.002444 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:39 crc kubenswrapper[4872]: E1009 08:20:39.011020 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ptfxz" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" Oct 09 08:20:39 crc kubenswrapper[4872]: E1009 08:20:39.906210 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 09 08:20:39 crc kubenswrapper[4872]: E1009 08:20:39.907284 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kbjsw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-dhnbs_openshift-marketplace(c4f8212a-5309-4088-b643-4f3d421e8799): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 09 08:20:39 crc kubenswrapper[4872]: E1009 08:20:39.908658 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-dhnbs" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" Oct 09 08:20:40 crc kubenswrapper[4872]: I1009 08:20:40.020847 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-8grb6" event={"ID":"326d4497-ac5f-4a64-afde-14ded51665b8","Type":"ContainerStarted","Data":"3fd043dc08defac869bd045be785698ef289a3c3fd4ba8716d860ab337ffd77b"} Oct 09 08:20:40 crc kubenswrapper[4872]: I1009 08:20:40.021678 4872 patch_prober.go:28] interesting pod/downloads-7954f5f757-nm6xv container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Oct 09 08:20:40 crc kubenswrapper[4872]: I1009 08:20:40.021768 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-nm6xv" podUID="192a1e23-c791-475f-be33-26f4b383a383" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.38:8080/\": dial tcp 10.217.0.38:8080: connect: connection refused" Oct 09 08:20:40 crc kubenswrapper[4872]: E1009 08:20:40.024283 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-dhnbs" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" Oct 09 08:20:40 crc kubenswrapper[4872]: I1009 08:20:40.051182 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-8grb6" podStartSLOduration=179.051158232 podStartE2EDuration="2m59.051158232s" podCreationTimestamp="2025-10-09 08:17:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:20:40.038655421 +0000 UTC m=+198.229184087" watchObservedRunningTime="2025-10-09 08:20:40.051158232 +0000 UTC m=+198.241686858" Oct 09 08:20:44 crc kubenswrapper[4872]: I1009 08:20:44.295904 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-nm6xv" Oct 09 08:20:58 crc kubenswrapper[4872]: I1009 08:20:58.130137 4872 generic.go:334] "Generic (PLEG): container finished" podID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerID="b0e388de262ddde7df0f2d2385ddbcd8d16bcfce0b661f79bafaaf2c5147699b" exitCode=0 Oct 09 08:20:58 crc kubenswrapper[4872]: I1009 08:20:58.130209 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerDied","Data":"b0e388de262ddde7df0f2d2385ddbcd8d16bcfce0b661f79bafaaf2c5147699b"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.149565 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerID="6e9d104a8225390e1dbda64fe56905ee7458e9c497e3296c6f6d00a649c4e4bf" exitCode=0 Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.149686 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerDied","Data":"6e9d104a8225390e1dbda64fe56905ee7458e9c497e3296c6f6d00a649c4e4bf"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.153862 4872 generic.go:334] "Generic (PLEG): container finished" podID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerID="14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9" exitCode=0 Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.153951 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerDied","Data":"14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.155655 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerStarted","Data":"662fbc4e940261f95dfd8f60230b2f8375accece7f43d556c3422186d20959fb"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.157317 4872 generic.go:334] "Generic (PLEG): container finished" podID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerID="eb00a8f453fdc96bfb713e5100d48c1667c987ca9cab2696319b2049eb61c870" exitCode=0 Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.157392 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerDied","Data":"eb00a8f453fdc96bfb713e5100d48c1667c987ca9cab2696319b2049eb61c870"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.159145 4872 generic.go:334] "Generic (PLEG): container finished" podID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerID="d4b2f69b21af0457086f8e1873fb16f2b3018c22a68d9a51ae7f4bc501f5613b" exitCode=0 Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.159224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerDied","Data":"d4b2f69b21af0457086f8e1873fb16f2b3018c22a68d9a51ae7f4bc501f5613b"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.161531 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerStarted","Data":"dbd59a508fe61497559fd66a1411437a963690037423678672b4c688a0f30f45"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.181926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerStarted","Data":"49872f1cde7aa3bb4481b301c2b8126b8a31625978f944e8bdb3ad600ea901c5"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.187982 4872 generic.go:334] "Generic (PLEG): container finished" podID="c4f8212a-5309-4088-b643-4f3d421e8799" containerID="c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622" exitCode=0 Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.188036 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerDied","Data":"c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622"} Oct 09 08:21:01 crc kubenswrapper[4872]: I1009 08:21:01.299179 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jds9b" podStartSLOduration=3.604703844 podStartE2EDuration="1m12.299158841s" podCreationTimestamp="2025-10-09 08:19:49 +0000 UTC" firstStartedPulling="2025-10-09 08:19:51.487738926 +0000 UTC m=+149.678267552" lastFinishedPulling="2025-10-09 08:21:00.182193923 +0000 UTC m=+218.372722549" observedRunningTime="2025-10-09 08:21:01.298876333 +0000 UTC m=+219.489404969" watchObservedRunningTime="2025-10-09 08:21:01.299158841 +0000 UTC m=+219.489687467" Oct 09 08:21:02 crc kubenswrapper[4872]: I1009 08:21:02.195206 4872 generic.go:334] "Generic (PLEG): container finished" podID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerID="662fbc4e940261f95dfd8f60230b2f8375accece7f43d556c3422186d20959fb" exitCode=0 Oct 09 08:21:02 crc kubenswrapper[4872]: I1009 08:21:02.195267 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerDied","Data":"662fbc4e940261f95dfd8f60230b2f8375accece7f43d556c3422186d20959fb"} Oct 09 08:21:02 crc kubenswrapper[4872]: I1009 08:21:02.197429 4872 generic.go:334] "Generic (PLEG): container finished" podID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerID="dbd59a508fe61497559fd66a1411437a963690037423678672b4c688a0f30f45" exitCode=0 Oct 09 08:21:02 crc kubenswrapper[4872]: I1009 08:21:02.197462 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerDied","Data":"dbd59a508fe61497559fd66a1411437a963690037423678672b4c688a0f30f45"} Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.204460 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerStarted","Data":"504b9c11a569113407147eb192419b4ea20bed14872de3a23c8f0aac5229ccf0"} Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.209099 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerStarted","Data":"ae272bd28a538fd4b37104ad47d2724ee06d2ae1ced1f933252665981644cc9d"} Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.210708 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerStarted","Data":"63f45fb1bda098e97f988a1fa026e0b1986712fa79a4d82799832e9be013b011"} Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.221409 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-blvp8" podStartSLOduration=3.129485855 podStartE2EDuration="1m13.221389608s" podCreationTimestamp="2025-10-09 08:19:50 +0000 UTC" firstStartedPulling="2025-10-09 08:19:52.512950158 +0000 UTC m=+150.703478784" lastFinishedPulling="2025-10-09 08:21:02.604853911 +0000 UTC m=+220.795382537" observedRunningTime="2025-10-09 08:21:03.218725527 +0000 UTC m=+221.409254173" watchObservedRunningTime="2025-10-09 08:21:03.221389608 +0000 UTC m=+221.411918244" Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.255984 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ptfxz" podStartSLOduration=3.075139178 podStartE2EDuration="1m16.255965891s" podCreationTimestamp="2025-10-09 08:19:47 +0000 UTC" firstStartedPulling="2025-10-09 08:19:49.194492493 +0000 UTC m=+147.385021119" lastFinishedPulling="2025-10-09 08:21:02.375319206 +0000 UTC m=+220.565847832" observedRunningTime="2025-10-09 08:21:03.235422455 +0000 UTC m=+221.425951091" watchObservedRunningTime="2025-10-09 08:21:03.255965891 +0000 UTC m=+221.446494517" Oct 09 08:21:03 crc kubenswrapper[4872]: I1009 08:21:03.256562 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-w6xb5" podStartSLOduration=2.781229961 podStartE2EDuration="1m16.256556929s" podCreationTimestamp="2025-10-09 08:19:47 +0000 UTC" firstStartedPulling="2025-10-09 08:19:49.175730881 +0000 UTC m=+147.366259507" lastFinishedPulling="2025-10-09 08:21:02.651057849 +0000 UTC m=+220.841586475" observedRunningTime="2025-10-09 08:21:03.251712082 +0000 UTC m=+221.442240718" watchObservedRunningTime="2025-10-09 08:21:03.256556929 +0000 UTC m=+221.447085555" Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.218284 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerStarted","Data":"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565"} Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.221654 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerStarted","Data":"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea"} Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.224351 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerStarted","Data":"25eab42144a0fd46352e059147c5bc6f5565cc588505a503594c6ef691b790aa"} Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.226942 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerStarted","Data":"25ea6628c78a1449dc02401e85c52ad82b9541ec4651364a3379c82a9962cab6"} Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.249209 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dhnbs" podStartSLOduration=2.544252756 podStartE2EDuration="1m16.249193699s" podCreationTimestamp="2025-10-09 08:19:48 +0000 UTC" firstStartedPulling="2025-10-09 08:19:50.27072155 +0000 UTC m=+148.461250176" lastFinishedPulling="2025-10-09 08:21:03.975662493 +0000 UTC m=+222.166191119" observedRunningTime="2025-10-09 08:21:04.245846797 +0000 UTC m=+222.436375433" watchObservedRunningTime="2025-10-09 08:21:04.249193699 +0000 UTC m=+222.439722325" Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.263873 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vz87x" podStartSLOduration=1.9087859310000002 podStartE2EDuration="1m13.263855976s" podCreationTimestamp="2025-10-09 08:19:51 +0000 UTC" firstStartedPulling="2025-10-09 08:19:52.529780961 +0000 UTC m=+150.720309587" lastFinishedPulling="2025-10-09 08:21:03.884851006 +0000 UTC m=+222.075379632" observedRunningTime="2025-10-09 08:21:04.26369025 +0000 UTC m=+222.454218876" watchObservedRunningTime="2025-10-09 08:21:04.263855976 +0000 UTC m=+222.454384602" Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.287428 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5w56c" podStartSLOduration=3.4630947819999998 podStartE2EDuration="1m14.287406173s" podCreationTimestamp="2025-10-09 08:19:50 +0000 UTC" firstStartedPulling="2025-10-09 08:19:52.517560518 +0000 UTC m=+150.708089144" lastFinishedPulling="2025-10-09 08:21:03.341871909 +0000 UTC m=+221.532400535" observedRunningTime="2025-10-09 08:21:04.283165744 +0000 UTC m=+222.473694370" watchObservedRunningTime="2025-10-09 08:21:04.287406173 +0000 UTC m=+222.477934809" Oct 09 08:21:04 crc kubenswrapper[4872]: I1009 08:21:04.308743 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lvbdw" podStartSLOduration=3.721881108 podStartE2EDuration="1m17.308725273s" podCreationTimestamp="2025-10-09 08:19:47 +0000 UTC" firstStartedPulling="2025-10-09 08:19:50.340103934 +0000 UTC m=+148.530632560" lastFinishedPulling="2025-10-09 08:21:03.926948099 +0000 UTC m=+222.117476725" observedRunningTime="2025-10-09 08:21:04.307152055 +0000 UTC m=+222.497680701" watchObservedRunningTime="2025-10-09 08:21:04.308725273 +0000 UTC m=+222.499253909" Oct 09 08:21:07 crc kubenswrapper[4872]: I1009 08:21:07.830008 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:21:07 crc kubenswrapper[4872]: I1009 08:21:07.830078 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.049145 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.050067 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.075186 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.093208 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.157157 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.157214 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.157255 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.157824 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.157889 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85" gracePeriod=600 Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.287895 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.292484 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.294885 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.295008 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.352456 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.442790 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.442874 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:08 crc kubenswrapper[4872]: I1009 08:21:08.486779 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:09 crc kubenswrapper[4872]: I1009 08:21:09.253534 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85" exitCode=0 Oct 09 08:21:09 crc kubenswrapper[4872]: I1009 08:21:09.253629 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85"} Oct 09 08:21:09 crc kubenswrapper[4872]: I1009 08:21:09.254908 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f"} Oct 09 08:21:09 crc kubenswrapper[4872]: I1009 08:21:09.302029 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:09 crc kubenswrapper[4872]: I1009 08:21:09.303547 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.049343 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.049701 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.097974 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.303697 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.391007 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.391075 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.458561 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.981172 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:21:10 crc kubenswrapper[4872]: I1009 08:21:10.982194 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.050956 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.303845 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.332455 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.379513 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.395715 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.395775 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.438539 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.580593 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.581952 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dhnbs" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="registry-server" containerID="cri-o://95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565" gracePeriod=2 Oct 09 08:21:11 crc kubenswrapper[4872]: I1009 08:21:11.957959 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.042845 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbjsw\" (UniqueName: \"kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw\") pod \"c4f8212a-5309-4088-b643-4f3d421e8799\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.042959 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content\") pod \"c4f8212a-5309-4088-b643-4f3d421e8799\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.044082 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities" (OuterVolumeSpecName: "utilities") pod "c4f8212a-5309-4088-b643-4f3d421e8799" (UID: "c4f8212a-5309-4088-b643-4f3d421e8799"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.048613 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw" (OuterVolumeSpecName: "kube-api-access-kbjsw") pod "c4f8212a-5309-4088-b643-4f3d421e8799" (UID: "c4f8212a-5309-4088-b643-4f3d421e8799"). InnerVolumeSpecName "kube-api-access-kbjsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.043157 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities\") pod \"c4f8212a-5309-4088-b643-4f3d421e8799\" (UID: \"c4f8212a-5309-4088-b643-4f3d421e8799\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.065220 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbjsw\" (UniqueName: \"kubernetes.io/projected/c4f8212a-5309-4088-b643-4f3d421e8799-kube-api-access-kbjsw\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.065258 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.110432 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4f8212a-5309-4088-b643-4f3d421e8799" (UID: "c4f8212a-5309-4088-b643-4f3d421e8799"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.166074 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4f8212a-5309-4088-b643-4f3d421e8799-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.274065 4872 generic.go:334] "Generic (PLEG): container finished" podID="c4f8212a-5309-4088-b643-4f3d421e8799" containerID="95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565" exitCode=0 Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.274247 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerDied","Data":"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565"} Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.274283 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhnbs" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.274305 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhnbs" event={"ID":"c4f8212a-5309-4088-b643-4f3d421e8799","Type":"ContainerDied","Data":"a8e47bce071f2cf5bdf5cc4167d283ef71681300b1bce37a68f0034c911699fc"} Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.274326 4872 scope.go:117] "RemoveContainer" containerID="95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.276406 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lvbdw" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="registry-server" containerID="cri-o://ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea" gracePeriod=2 Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.292320 4872 scope.go:117] "RemoveContainer" containerID="c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.308152 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.311825 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dhnbs"] Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.329118 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.334184 4872 scope.go:117] "RemoveContainer" containerID="dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.395689 4872 scope.go:117] "RemoveContainer" containerID="95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565" Oct 09 08:21:12 crc kubenswrapper[4872]: E1009 08:21:12.396210 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565\": container with ID starting with 95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565 not found: ID does not exist" containerID="95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.396257 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565"} err="failed to get container status \"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565\": rpc error: code = NotFound desc = could not find container \"95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565\": container with ID starting with 95c4e2c70cd3c5f458bf49e3e8098018b96c29ab3b69ffa393befb6f0b01b565 not found: ID does not exist" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.396287 4872 scope.go:117] "RemoveContainer" containerID="c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622" Oct 09 08:21:12 crc kubenswrapper[4872]: E1009 08:21:12.396870 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622\": container with ID starting with c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622 not found: ID does not exist" containerID="c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.396914 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622"} err="failed to get container status \"c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622\": rpc error: code = NotFound desc = could not find container \"c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622\": container with ID starting with c966c6c09c2c52f8c46219fcf51ef31d003f44fa978eb5b30eb6ffc74c011622 not found: ID does not exist" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.396940 4872 scope.go:117] "RemoveContainer" containerID="dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61" Oct 09 08:21:12 crc kubenswrapper[4872]: E1009 08:21:12.397543 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61\": container with ID starting with dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61 not found: ID does not exist" containerID="dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.397574 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61"} err="failed to get container status \"dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61\": rpc error: code = NotFound desc = could not find container \"dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61\": container with ID starting with dfa57773c40ccbcdbca5700122b5108d2f1179be88755c2b26e572d8ba15cd61 not found: ID does not exist" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.469571 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" path="/var/lib/kubelet/pods/c4f8212a-5309-4088-b643-4f3d421e8799/volumes" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.605390 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.772591 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5k27\" (UniqueName: \"kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27\") pod \"71f030ad-a0a1-49b8-9910-1bf39006e56a\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.772685 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities\") pod \"71f030ad-a0a1-49b8-9910-1bf39006e56a\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.772799 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content\") pod \"71f030ad-a0a1-49b8-9910-1bf39006e56a\" (UID: \"71f030ad-a0a1-49b8-9910-1bf39006e56a\") " Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.788102 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities" (OuterVolumeSpecName: "utilities") pod "71f030ad-a0a1-49b8-9910-1bf39006e56a" (UID: "71f030ad-a0a1-49b8-9910-1bf39006e56a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.788355 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27" (OuterVolumeSpecName: "kube-api-access-g5k27") pod "71f030ad-a0a1-49b8-9910-1bf39006e56a" (UID: "71f030ad-a0a1-49b8-9910-1bf39006e56a"). InnerVolumeSpecName "kube-api-access-g5k27". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.828901 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71f030ad-a0a1-49b8-9910-1bf39006e56a" (UID: "71f030ad-a0a1-49b8-9910-1bf39006e56a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.874726 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5k27\" (UniqueName: \"kubernetes.io/projected/71f030ad-a0a1-49b8-9910-1bf39006e56a-kube-api-access-g5k27\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.874769 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:12 crc kubenswrapper[4872]: I1009 08:21:12.874780 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71f030ad-a0a1-49b8-9910-1bf39006e56a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.286721 4872 generic.go:334] "Generic (PLEG): container finished" podID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerID="ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea" exitCode=0 Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.286805 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lvbdw" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.286802 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerDied","Data":"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea"} Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.292617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lvbdw" event={"ID":"71f030ad-a0a1-49b8-9910-1bf39006e56a","Type":"ContainerDied","Data":"b9707481ebb6894284443d9e5f862c48cc89969c6f02e19026e9465f4d68c43c"} Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.292692 4872 scope.go:117] "RemoveContainer" containerID="ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.323746 4872 scope.go:117] "RemoveContainer" containerID="14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.335713 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.341572 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lvbdw"] Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.359362 4872 scope.go:117] "RemoveContainer" containerID="1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.384729 4872 scope.go:117] "RemoveContainer" containerID="ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea" Oct 09 08:21:13 crc kubenswrapper[4872]: E1009 08:21:13.388717 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea\": container with ID starting with ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea not found: ID does not exist" containerID="ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.388748 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea"} err="failed to get container status \"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea\": rpc error: code = NotFound desc = could not find container \"ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea\": container with ID starting with ca3eea40a4c759a6df266b7afe0277730dfd7ee56e06ff295c34b8fc87cb5dea not found: ID does not exist" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.388770 4872 scope.go:117] "RemoveContainer" containerID="14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9" Oct 09 08:21:13 crc kubenswrapper[4872]: E1009 08:21:13.389132 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9\": container with ID starting with 14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9 not found: ID does not exist" containerID="14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.389233 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9"} err="failed to get container status \"14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9\": rpc error: code = NotFound desc = could not find container \"14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9\": container with ID starting with 14580669c97d8c071e7431f151ba441eccc5fad37b2b2add50171d8be5c316f9 not found: ID does not exist" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.389312 4872 scope.go:117] "RemoveContainer" containerID="1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061" Oct 09 08:21:13 crc kubenswrapper[4872]: E1009 08:21:13.389690 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061\": container with ID starting with 1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061 not found: ID does not exist" containerID="1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.389714 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061"} err="failed to get container status \"1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061\": rpc error: code = NotFound desc = could not find container \"1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061\": container with ID starting with 1bf036889c74ccf2767135b0e87fccab71a6fb35b3e185132d89b45a460bd061 not found: ID does not exist" Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.786410 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.787034 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-blvp8" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="registry-server" containerID="cri-o://504b9c11a569113407147eb192419b4ea20bed14872de3a23c8f0aac5229ccf0" gracePeriod=2 Oct 09 08:21:13 crc kubenswrapper[4872]: I1009 08:21:13.980403 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:21:14 crc kubenswrapper[4872]: I1009 08:21:14.301130 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vz87x" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="registry-server" containerID="cri-o://25ea6628c78a1449dc02401e85c52ad82b9541ec4651364a3379c82a9962cab6" gracePeriod=2 Oct 09 08:21:14 crc kubenswrapper[4872]: I1009 08:21:14.468119 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" path="/var/lib/kubelet/pods/71f030ad-a0a1-49b8-9910-1bf39006e56a/volumes" Oct 09 08:21:15 crc kubenswrapper[4872]: I1009 08:21:15.311651 4872 generic.go:334] "Generic (PLEG): container finished" podID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerID="504b9c11a569113407147eb192419b4ea20bed14872de3a23c8f0aac5229ccf0" exitCode=0 Oct 09 08:21:15 crc kubenswrapper[4872]: I1009 08:21:15.311767 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerDied","Data":"504b9c11a569113407147eb192419b4ea20bed14872de3a23c8f0aac5229ccf0"} Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.149495 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.216497 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkzr8\" (UniqueName: \"kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8\") pod \"a5c8e529-78d6-4132-b94e-91df9aadda37\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.216962 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content\") pod \"a5c8e529-78d6-4132-b94e-91df9aadda37\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.216999 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities\") pod \"a5c8e529-78d6-4132-b94e-91df9aadda37\" (UID: \"a5c8e529-78d6-4132-b94e-91df9aadda37\") " Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.220555 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities" (OuterVolumeSpecName: "utilities") pod "a5c8e529-78d6-4132-b94e-91df9aadda37" (UID: "a5c8e529-78d6-4132-b94e-91df9aadda37"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.222573 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8" (OuterVolumeSpecName: "kube-api-access-hkzr8") pod "a5c8e529-78d6-4132-b94e-91df9aadda37" (UID: "a5c8e529-78d6-4132-b94e-91df9aadda37"). InnerVolumeSpecName "kube-api-access-hkzr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.246951 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a5c8e529-78d6-4132-b94e-91df9aadda37" (UID: "a5c8e529-78d6-4132-b94e-91df9aadda37"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.317810 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkzr8\" (UniqueName: \"kubernetes.io/projected/a5c8e529-78d6-4132-b94e-91df9aadda37-kube-api-access-hkzr8\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.317836 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.317846 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a5c8e529-78d6-4132-b94e-91df9aadda37-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.318815 4872 generic.go:334] "Generic (PLEG): container finished" podID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerID="25ea6628c78a1449dc02401e85c52ad82b9541ec4651364a3379c82a9962cab6" exitCode=0 Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.318884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerDied","Data":"25ea6628c78a1449dc02401e85c52ad82b9541ec4651364a3379c82a9962cab6"} Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.320894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-blvp8" event={"ID":"a5c8e529-78d6-4132-b94e-91df9aadda37","Type":"ContainerDied","Data":"0931e919ea5a7af76bfe120bbe5a3c75fe57e69bb55e3285f6f133a4e81b0350"} Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.320939 4872 scope.go:117] "RemoveContainer" containerID="504b9c11a569113407147eb192419b4ea20bed14872de3a23c8f0aac5229ccf0" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.320942 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-blvp8" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.342376 4872 scope.go:117] "RemoveContainer" containerID="6e9d104a8225390e1dbda64fe56905ee7458e9c497e3296c6f6d00a649c4e4bf" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.348447 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.351310 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-blvp8"] Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.379853 4872 scope.go:117] "RemoveContainer" containerID="922e79b9edbb83e33796b0dd57f6f8aaba5d2f182a9c3164cb8a0d3ffe6174e9" Oct 09 08:21:16 crc kubenswrapper[4872]: I1009 08:21:16.471805 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" path="/var/lib/kubelet/pods/a5c8e529-78d6-4132-b94e-91df9aadda37/volumes" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.049395 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.132144 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5bg6\" (UniqueName: \"kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6\") pod \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.132474 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content\") pod \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.132526 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities\") pod \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\" (UID: \"0e121152-fd8c-42a6-98a5-d7ea05dfab3e\") " Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.134313 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities" (OuterVolumeSpecName: "utilities") pod "0e121152-fd8c-42a6-98a5-d7ea05dfab3e" (UID: "0e121152-fd8c-42a6-98a5-d7ea05dfab3e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.135922 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6" (OuterVolumeSpecName: "kube-api-access-f5bg6") pod "0e121152-fd8c-42a6-98a5-d7ea05dfab3e" (UID: "0e121152-fd8c-42a6-98a5-d7ea05dfab3e"). InnerVolumeSpecName "kube-api-access-f5bg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.234114 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.234199 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5bg6\" (UniqueName: \"kubernetes.io/projected/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-kube-api-access-f5bg6\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.329818 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vz87x" event={"ID":"0e121152-fd8c-42a6-98a5-d7ea05dfab3e","Type":"ContainerDied","Data":"0f88be4b07c82b5ca709b3636d817b9938f9cdf39160bc2273730f9d695360c1"} Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.329868 4872 scope.go:117] "RemoveContainer" containerID="25ea6628c78a1449dc02401e85c52ad82b9541ec4651364a3379c82a9962cab6" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.329910 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vz87x" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.345677 4872 scope.go:117] "RemoveContainer" containerID="dbd59a508fe61497559fd66a1411437a963690037423678672b4c688a0f30f45" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.362129 4872 scope.go:117] "RemoveContainer" containerID="4fb50b9d16a3ce8520609c47d06d6c51132368670bbba8aac8a1cba97cdfa7f9" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.533048 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0e121152-fd8c-42a6-98a5-d7ea05dfab3e" (UID: "0e121152-fd8c-42a6-98a5-d7ea05dfab3e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.539128 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0e121152-fd8c-42a6-98a5-d7ea05dfab3e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.657689 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:21:17 crc kubenswrapper[4872]: I1009 08:21:17.669977 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vz87x"] Oct 09 08:21:18 crc kubenswrapper[4872]: I1009 08:21:18.467393 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" path="/var/lib/kubelet/pods/0e121152-fd8c-42a6-98a5-d7ea05dfab3e/volumes" Oct 09 08:21:32 crc kubenswrapper[4872]: I1009 08:21:32.700168 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:21:57 crc kubenswrapper[4872]: I1009 08:21:57.723573 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" containerID="cri-o://802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c" gracePeriod=15 Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.140764 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.169884 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-848fcd468b-gpn6m"] Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170094 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170109 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170121 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170127 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170138 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170146 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170153 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170159 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170169 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170175 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170184 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170189 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170196 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170202 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="extract-utilities" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170213 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54fadc33-526c-4803-a659-90cac0c4a861" containerName="pruner" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170219 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="54fadc33-526c-4803-a659-90cac0c4a861" containerName="pruner" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170229 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170234 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170241 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170246 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170253 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170258 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170266 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170272 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170279 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170284 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="extract-content" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.170293 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170299 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170378 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="54fadc33-526c-4803-a659-90cac0c4a861" containerName="pruner" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170393 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e121152-fd8c-42a6-98a5-d7ea05dfab3e" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170399 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="75af6421-130b-4287-a014-d17512f708e4" containerName="oauth-openshift" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170408 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f8212a-5309-4088-b643-4f3d421e8799" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170417 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5c8e529-78d6-4132-b94e-91df9aadda37" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170424 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="71f030ad-a0a1-49b8-9910-1bf39006e56a" containerName="registry-server" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.170788 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.195093 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-848fcd468b-gpn6m"] Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336421 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336485 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336504 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8bflb\" (UniqueName: \"kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336526 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336552 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336578 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336605 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336622 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336663 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336683 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336699 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336725 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336745 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336764 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert\") pod \"75af6421-130b-4287-a014-d17512f708e4\" (UID: \"75af6421-130b-4287-a014-d17512f708e4\") " Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336933 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9rvg\" (UniqueName: \"kubernetes.io/projected/22cba097-5696-4eba-b50f-4e3ecc275a81-kube-api-access-h9rvg\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336963 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.336988 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-service-ca\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337022 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-error\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337054 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-policies\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337069 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337085 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337101 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-login\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337126 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-router-certs\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337154 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-dir\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337173 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-session\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337208 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337237 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.337736 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.338038 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.338249 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.338313 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.338655 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.343072 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb" (OuterVolumeSpecName: "kube-api-access-8bflb") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "kube-api-access-8bflb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.344820 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.351339 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.351977 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.353801 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.354069 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.355398 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.355501 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.355742 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "75af6421-130b-4287-a014-d17512f708e4" (UID: "75af6421-130b-4287-a014-d17512f708e4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438138 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438440 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9rvg\" (UniqueName: \"kubernetes.io/projected/22cba097-5696-4eba-b50f-4e3ecc275a81-kube-api-access-h9rvg\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438537 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438655 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-service-ca\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438784 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-error\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.438906 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-policies\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439011 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439109 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-login\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439200 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439300 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-router-certs\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439399 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-dir\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439489 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439588 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-session\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439704 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439784 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-dir\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439327 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-service-ca\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.439455 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-cliconfig\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440382 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440456 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440480 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440495 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440508 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8bflb\" (UniqueName: \"kubernetes.io/projected/75af6421-130b-4287-a014-d17512f708e4-kube-api-access-8bflb\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440522 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440537 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440566 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440590 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440602 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440614 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440628 4872 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/75af6421-130b-4287-a014-d17512f708e4-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440657 4872 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440670 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.440681 4872 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/75af6421-130b-4287-a014-d17512f708e4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.441220 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-error\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.441225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.443803 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/22cba097-5696-4eba-b50f-4e3ecc275a81-audit-policies\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.444834 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-router-certs\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.446629 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.450713 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-template-login\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.451042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.451178 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-session\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.451319 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/22cba097-5696-4eba-b50f-4e3ecc275a81-v4-0-config-system-serving-cert\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.458220 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9rvg\" (UniqueName: \"kubernetes.io/projected/22cba097-5696-4eba-b50f-4e3ecc275a81-kube-api-access-h9rvg\") pod \"oauth-openshift-848fcd468b-gpn6m\" (UID: \"22cba097-5696-4eba-b50f-4e3ecc275a81\") " pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.491983 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.554566 4872 generic.go:334] "Generic (PLEG): container finished" podID="75af6421-130b-4287-a014-d17512f708e4" containerID="802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c" exitCode=0 Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.554620 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" event={"ID":"75af6421-130b-4287-a014-d17512f708e4","Type":"ContainerDied","Data":"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c"} Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.554668 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" event={"ID":"75af6421-130b-4287-a014-d17512f708e4","Type":"ContainerDied","Data":"9d954d2c5cde8f3daf6ede45b7282d0ed76e396d6a4a86f5596d3bf6710ed3c6"} Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.554691 4872 scope.go:117] "RemoveContainer" containerID="802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.554838 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g747n" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.571740 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.574207 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g747n"] Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.581102 4872 scope.go:117] "RemoveContainer" containerID="802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c" Oct 09 08:21:58 crc kubenswrapper[4872]: E1009 08:21:58.581534 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c\": container with ID starting with 802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c not found: ID does not exist" containerID="802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.581592 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c"} err="failed to get container status \"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c\": rpc error: code = NotFound desc = could not find container \"802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c\": container with ID starting with 802920e17e5aa6e20a94d5c604f9af145db48a0f5041d4e4b1078bbc2fc7764c not found: ID does not exist" Oct 09 08:21:58 crc kubenswrapper[4872]: I1009 08:21:58.696221 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-848fcd468b-gpn6m"] Oct 09 08:21:59 crc kubenswrapper[4872]: I1009 08:21:59.560976 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" event={"ID":"22cba097-5696-4eba-b50f-4e3ecc275a81","Type":"ContainerStarted","Data":"742103dee5d9846cca764fe34650fc85bc1b679949e24aab1ca6a7722efd30d6"} Oct 09 08:21:59 crc kubenswrapper[4872]: I1009 08:21:59.561560 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:59 crc kubenswrapper[4872]: I1009 08:21:59.561576 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" event={"ID":"22cba097-5696-4eba-b50f-4e3ecc275a81","Type":"ContainerStarted","Data":"121a9936173e178579572404ed9e6585ef54ef9c4c2c11ba3ea3b3dcc9c5f042"} Oct 09 08:21:59 crc kubenswrapper[4872]: I1009 08:21:59.567343 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" Oct 09 08:21:59 crc kubenswrapper[4872]: I1009 08:21:59.584344 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-848fcd468b-gpn6m" podStartSLOduration=27.584327001 podStartE2EDuration="27.584327001s" podCreationTimestamp="2025-10-09 08:21:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:21:59.583395484 +0000 UTC m=+277.773924110" watchObservedRunningTime="2025-10-09 08:21:59.584327001 +0000 UTC m=+277.774855627" Oct 09 08:22:00 crc kubenswrapper[4872]: I1009 08:22:00.472210 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75af6421-130b-4287-a014-d17512f708e4" path="/var/lib/kubelet/pods/75af6421-130b-4287-a014-d17512f708e4/volumes" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.215522 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.216489 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-w6xb5" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="registry-server" containerID="cri-o://ae272bd28a538fd4b37104ad47d2724ee06d2ae1ced1f933252665981644cc9d" gracePeriod=30 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.220213 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.220422 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ptfxz" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="registry-server" containerID="cri-o://63f45fb1bda098e97f988a1fa026e0b1986712fa79a4d82799832e9be013b011" gracePeriod=30 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.236021 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.236241 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" containerID="cri-o://2e2d171e335a67b9100fa1dab0e810e1e661ef0b2ca2f70f5ec37eec6a35da65" gracePeriod=30 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.253184 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.253436 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jds9b" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="registry-server" containerID="cri-o://49872f1cde7aa3bb4481b301c2b8126b8a31625978f944e8bdb3ad600ea901c5" gracePeriod=30 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.262262 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69jm8"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.263058 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.268882 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.269125 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5w56c" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="registry-server" containerID="cri-o://25eab42144a0fd46352e059147c5bc6f5565cc588505a503594c6ef691b790aa" gracePeriod=30 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.277813 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69jm8"] Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.413290 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.413910 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlcws\" (UniqueName: \"kubernetes.io/projected/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-kube-api-access-rlcws\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.413950 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.515375 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlcws\" (UniqueName: \"kubernetes.io/projected/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-kube-api-access-rlcws\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.515430 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.515490 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.517603 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.521792 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.532615 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlcws\" (UniqueName: \"kubernetes.io/projected/6b2d20f2-e34f-4b19-85fb-0e0bfffe849d-kube-api-access-rlcws\") pod \"marketplace-operator-79b997595-69jm8\" (UID: \"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d\") " pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.592259 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.627546 4872 generic.go:334] "Generic (PLEG): container finished" podID="a18c8409-0736-42ad-bb89-717d51c1c767" containerID="2e2d171e335a67b9100fa1dab0e810e1e661ef0b2ca2f70f5ec37eec6a35da65" exitCode=0 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.627603 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" event={"ID":"a18c8409-0736-42ad-bb89-717d51c1c767","Type":"ContainerDied","Data":"2e2d171e335a67b9100fa1dab0e810e1e661ef0b2ca2f70f5ec37eec6a35da65"} Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.629266 4872 generic.go:334] "Generic (PLEG): container finished" podID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerID="63f45fb1bda098e97f988a1fa026e0b1986712fa79a4d82799832e9be013b011" exitCode=0 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.629321 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerDied","Data":"63f45fb1bda098e97f988a1fa026e0b1986712fa79a4d82799832e9be013b011"} Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.631049 4872 generic.go:334] "Generic (PLEG): container finished" podID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerID="49872f1cde7aa3bb4481b301c2b8126b8a31625978f944e8bdb3ad600ea901c5" exitCode=0 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.631093 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerDied","Data":"49872f1cde7aa3bb4481b301c2b8126b8a31625978f944e8bdb3ad600ea901c5"} Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.634495 4872 generic.go:334] "Generic (PLEG): container finished" podID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerID="25eab42144a0fd46352e059147c5bc6f5565cc588505a503594c6ef691b790aa" exitCode=0 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.634560 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerDied","Data":"25eab42144a0fd46352e059147c5bc6f5565cc588505a503594c6ef691b790aa"} Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.638109 4872 generic.go:334] "Generic (PLEG): container finished" podID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerID="ae272bd28a538fd4b37104ad47d2724ee06d2ae1ced1f933252665981644cc9d" exitCode=0 Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.638145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerDied","Data":"ae272bd28a538fd4b37104ad47d2724ee06d2ae1ced1f933252665981644cc9d"} Oct 09 08:22:12 crc kubenswrapper[4872]: I1009 08:22:12.791586 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-69jm8"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.030357 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.111139 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.113250 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.121220 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124430 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities\") pod \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124474 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics\") pod \"a18c8409-0736-42ad-bb89-717d51c1c767\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124493 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wbdv\" (UniqueName: \"kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv\") pod \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124517 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bcbt\" (UniqueName: \"kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt\") pod \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124549 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content\") pod \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124564 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content\") pod \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\" (UID: \"034eb0f7-581e-4e3d-aa86-12d7dc73690a\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca\") pod \"a18c8409-0736-42ad-bb89-717d51c1c767\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124602 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzfmh\" (UniqueName: \"kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh\") pod \"a18c8409-0736-42ad-bb89-717d51c1c767\" (UID: \"a18c8409-0736-42ad-bb89-717d51c1c767\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124623 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities\") pod \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\" (UID: \"8839fa4b-4de5-4399-a83d-4c3cced15cd5\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124652 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content\") pod \"933f1f0c-fc00-4ca8-b364-d97cebb40502\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124681 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh7bc\" (UniqueName: \"kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc\") pod \"933f1f0c-fc00-4ca8-b364-d97cebb40502\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.124697 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities\") pod \"933f1f0c-fc00-4ca8-b364-d97cebb40502\" (UID: \"933f1f0c-fc00-4ca8-b364-d97cebb40502\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.125467 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities" (OuterVolumeSpecName: "utilities") pod "933f1f0c-fc00-4ca8-b364-d97cebb40502" (UID: "933f1f0c-fc00-4ca8-b364-d97cebb40502"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.125502 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities" (OuterVolumeSpecName: "utilities") pod "034eb0f7-581e-4e3d-aa86-12d7dc73690a" (UID: "034eb0f7-581e-4e3d-aa86-12d7dc73690a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.127849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities" (OuterVolumeSpecName: "utilities") pod "8839fa4b-4de5-4399-a83d-4c3cced15cd5" (UID: "8839fa4b-4de5-4399-a83d-4c3cced15cd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.131118 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "a18c8409-0736-42ad-bb89-717d51c1c767" (UID: "a18c8409-0736-42ad-bb89-717d51c1c767"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.138438 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh" (OuterVolumeSpecName: "kube-api-access-fzfmh") pod "a18c8409-0736-42ad-bb89-717d51c1c767" (UID: "a18c8409-0736-42ad-bb89-717d51c1c767"). InnerVolumeSpecName "kube-api-access-fzfmh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.139696 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv" (OuterVolumeSpecName: "kube-api-access-5wbdv") pod "8839fa4b-4de5-4399-a83d-4c3cced15cd5" (UID: "8839fa4b-4de5-4399-a83d-4c3cced15cd5"). InnerVolumeSpecName "kube-api-access-5wbdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.147578 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt" (OuterVolumeSpecName: "kube-api-access-4bcbt") pod "034eb0f7-581e-4e3d-aa86-12d7dc73690a" (UID: "034eb0f7-581e-4e3d-aa86-12d7dc73690a"). InnerVolumeSpecName "kube-api-access-4bcbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.152463 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc" (OuterVolumeSpecName: "kube-api-access-hh7bc") pod "933f1f0c-fc00-4ca8-b364-d97cebb40502" (UID: "933f1f0c-fc00-4ca8-b364-d97cebb40502"). InnerVolumeSpecName "kube-api-access-hh7bc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.158298 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "a18c8409-0736-42ad-bb89-717d51c1c767" (UID: "a18c8409-0736-42ad-bb89-717d51c1c767"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.165783 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "034eb0f7-581e-4e3d-aa86-12d7dc73690a" (UID: "034eb0f7-581e-4e3d-aa86-12d7dc73690a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.182997 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.199377 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8839fa4b-4de5-4399-a83d-4c3cced15cd5" (UID: "8839fa4b-4de5-4399-a83d-4c3cced15cd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.211583 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "933f1f0c-fc00-4ca8-b364-d97cebb40502" (UID: "933f1f0c-fc00-4ca8-b364-d97cebb40502"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225659 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzfmh\" (UniqueName: \"kubernetes.io/projected/a18c8409-0736-42ad-bb89-717d51c1c767-kube-api-access-fzfmh\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225700 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225714 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225727 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh7bc\" (UniqueName: \"kubernetes.io/projected/933f1f0c-fc00-4ca8-b364-d97cebb40502-kube-api-access-hh7bc\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225742 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/933f1f0c-fc00-4ca8-b364-d97cebb40502-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225753 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225765 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225778 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wbdv\" (UniqueName: \"kubernetes.io/projected/8839fa4b-4de5-4399-a83d-4c3cced15cd5-kube-api-access-5wbdv\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225789 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bcbt\" (UniqueName: \"kubernetes.io/projected/034eb0f7-581e-4e3d-aa86-12d7dc73690a-kube-api-access-4bcbt\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225805 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8839fa4b-4de5-4399-a83d-4c3cced15cd5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225817 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/034eb0f7-581e-4e3d-aa86-12d7dc73690a-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.225828 4872 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a18c8409-0736-42ad-bb89-717d51c1c767-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.326440 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content\") pod \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.326799 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities\") pod \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.326852 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcfmq\" (UniqueName: \"kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq\") pod \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\" (UID: \"fc1ccd35-725f-453b-988e-0d2da9b17b6c\") " Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.328410 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities" (OuterVolumeSpecName: "utilities") pod "fc1ccd35-725f-453b-988e-0d2da9b17b6c" (UID: "fc1ccd35-725f-453b-988e-0d2da9b17b6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.330627 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq" (OuterVolumeSpecName: "kube-api-access-lcfmq") pod "fc1ccd35-725f-453b-988e-0d2da9b17b6c" (UID: "fc1ccd35-725f-453b-988e-0d2da9b17b6c"). InnerVolumeSpecName "kube-api-access-lcfmq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.427610 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fc1ccd35-725f-453b-988e-0d2da9b17b6c" (UID: "fc1ccd35-725f-453b-988e-0d2da9b17b6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.428213 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.428244 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fc1ccd35-725f-453b-988e-0d2da9b17b6c-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.428254 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcfmq\" (UniqueName: \"kubernetes.io/projected/fc1ccd35-725f-453b-988e-0d2da9b17b6c-kube-api-access-lcfmq\") on node \"crc\" DevicePath \"\"" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.644249 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.644261 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-xlstb" event={"ID":"a18c8409-0736-42ad-bb89-717d51c1c767","Type":"ContainerDied","Data":"65aff83636984223f74a79339971275fa92149e82884c71fd5a949a95f9fe51d"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.644329 4872 scope.go:117] "RemoveContainer" containerID="2e2d171e335a67b9100fa1dab0e810e1e661ef0b2ca2f70f5ec37eec6a35da65" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.646416 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" event={"ID":"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d","Type":"ContainerStarted","Data":"544044bd46d1a421bb58e93df1d554016e6562d1c7ff2867e679a99b31d38cb4"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.646439 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" event={"ID":"6b2d20f2-e34f-4b19-85fb-0e0bfffe849d","Type":"ContainerStarted","Data":"897c2fc018d35ab49fea520c6a5ca9a174ede44987cc08cec50c3a4fcd16192c"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.646791 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.649880 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jds9b" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.649873 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jds9b" event={"ID":"034eb0f7-581e-4e3d-aa86-12d7dc73690a","Type":"ContainerDied","Data":"b41545544afdccfb640576a1cf20ad2cba40486e076a74af31055c29ef440a24"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.654771 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5w56c" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.654883 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5w56c" event={"ID":"fc1ccd35-725f-453b-988e-0d2da9b17b6c","Type":"ContainerDied","Data":"1f93d2325f3feb05af21cfffc633091e0874adc9715753e4b62dbe4404f21957"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.655252 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.660892 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-w6xb5" event={"ID":"8839fa4b-4de5-4399-a83d-4c3cced15cd5","Type":"ContainerDied","Data":"852d683e3d62168542238cc114530f0e5f66b95f319e35679484e779a890c954"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.661837 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-w6xb5" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.667308 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ptfxz" event={"ID":"933f1f0c-fc00-4ca8-b364-d97cebb40502","Type":"ContainerDied","Data":"aa8860dd9434f67289f5161c31b0a3a6c3d28aa2bd3ca6d056d00f6763e266c4"} Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.667474 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ptfxz" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.669449 4872 scope.go:117] "RemoveContainer" containerID="49872f1cde7aa3bb4481b301c2b8126b8a31625978f944e8bdb3ad600ea901c5" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.686925 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-69jm8" podStartSLOduration=1.686900069 podStartE2EDuration="1.686900069s" podCreationTimestamp="2025-10-09 08:22:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:22:13.670384016 +0000 UTC m=+291.860912642" watchObservedRunningTime="2025-10-09 08:22:13.686900069 +0000 UTC m=+291.877428695" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.691335 4872 scope.go:117] "RemoveContainer" containerID="b0e388de262ddde7df0f2d2385ddbcd8d16bcfce0b661f79bafaaf2c5147699b" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.704330 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.708342 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-xlstb"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.744532 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.746166 4872 scope.go:117] "RemoveContainer" containerID="4fa4bfd3fce9ca84fb76d1d7799db694e6e0a2cf57102874ebd71d3b90ebfe6e" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.750959 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jds9b"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.759247 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.760537 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5w56c"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.763971 4872 scope.go:117] "RemoveContainer" containerID="25eab42144a0fd46352e059147c5bc6f5565cc588505a503594c6ef691b790aa" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.768024 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.772085 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ptfxz"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.785898 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.788689 4872 scope.go:117] "RemoveContainer" containerID="662fbc4e940261f95dfd8f60230b2f8375accece7f43d556c3422186d20959fb" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.790830 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-w6xb5"] Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.801297 4872 scope.go:117] "RemoveContainer" containerID="e15be488cd16c8db5e924abb709f534403245b1a7366bbedac3cb83f210d7d56" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.816399 4872 scope.go:117] "RemoveContainer" containerID="ae272bd28a538fd4b37104ad47d2724ee06d2ae1ced1f933252665981644cc9d" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.835618 4872 scope.go:117] "RemoveContainer" containerID="eb00a8f453fdc96bfb713e5100d48c1667c987ca9cab2696319b2049eb61c870" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.849732 4872 scope.go:117] "RemoveContainer" containerID="4454fc1f0fb1d18e60c99630f8ca6e3d35bfdaca2a48a7de07458ddd3f1b5d6d" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.862117 4872 scope.go:117] "RemoveContainer" containerID="63f45fb1bda098e97f988a1fa026e0b1986712fa79a4d82799832e9be013b011" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.873359 4872 scope.go:117] "RemoveContainer" containerID="d4b2f69b21af0457086f8e1873fb16f2b3018c22a68d9a51ae7f4bc501f5613b" Oct 09 08:22:13 crc kubenswrapper[4872]: I1009 08:22:13.885545 4872 scope.go:117] "RemoveContainer" containerID="61ef1e26f3519f47ed3ca5c70acc01f7d82f27e3599b4abf3dbad4d470d1d4d4" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433452 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433686 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433702 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433715 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433723 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433735 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433743 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433753 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433760 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433769 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433777 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433786 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433794 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433803 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433810 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433822 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433829 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="extract-utilities" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433839 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433848 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433861 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433869 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433879 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433889 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="extract-content" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433898 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433905 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: E1009 08:22:14.433916 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.433923 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434026 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" containerName="marketplace-operator" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434044 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434057 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434067 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434075 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" containerName="registry-server" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.434876 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.438716 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.442175 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.443495 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.443561 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ck5w\" (UniqueName: \"kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.443596 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.469685 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="034eb0f7-581e-4e3d-aa86-12d7dc73690a" path="/var/lib/kubelet/pods/034eb0f7-581e-4e3d-aa86-12d7dc73690a/volumes" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.470336 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8839fa4b-4de5-4399-a83d-4c3cced15cd5" path="/var/lib/kubelet/pods/8839fa4b-4de5-4399-a83d-4c3cced15cd5/volumes" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.471008 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="933f1f0c-fc00-4ca8-b364-d97cebb40502" path="/var/lib/kubelet/pods/933f1f0c-fc00-4ca8-b364-d97cebb40502/volumes" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.472112 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a18c8409-0736-42ad-bb89-717d51c1c767" path="/var/lib/kubelet/pods/a18c8409-0736-42ad-bb89-717d51c1c767/volumes" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.472571 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc1ccd35-725f-453b-988e-0d2da9b17b6c" path="/var/lib/kubelet/pods/fc1ccd35-725f-453b-988e-0d2da9b17b6c/volumes" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.544557 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ck5w\" (UniqueName: \"kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.544662 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.545014 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.545126 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.545711 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.564110 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ck5w\" (UniqueName: \"kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w\") pod \"certified-operators-mxtxg\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.635225 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x8nzf"] Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.636435 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.641476 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.651336 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8nzf"] Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.748202 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-catalog-content\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.748266 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx7xt\" (UniqueName: \"kubernetes.io/projected/f41d6540-7af0-4aa7-a36a-629e55c85a6f-kube-api-access-vx7xt\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.748299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-utilities\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.752117 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.849283 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-catalog-content\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.849337 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx7xt\" (UniqueName: \"kubernetes.io/projected/f41d6540-7af0-4aa7-a36a-629e55c85a6f-kube-api-access-vx7xt\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.849358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-utilities\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.850255 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-catalog-content\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.851255 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f41d6540-7af0-4aa7-a36a-629e55c85a6f-utilities\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.868417 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx7xt\" (UniqueName: \"kubernetes.io/projected/f41d6540-7af0-4aa7-a36a-629e55c85a6f-kube-api-access-vx7xt\") pod \"community-operators-x8nzf\" (UID: \"f41d6540-7af0-4aa7-a36a-629e55c85a6f\") " pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.937235 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:22:14 crc kubenswrapper[4872]: W1009 08:22:14.946148 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76a38bf4_890c_4f97_a34d_33cc8adbc882.slice/crio-094e230a91ee796a38f8b920ec9ab99bacf23906fbf2fa79ded8f87e45199fbc WatchSource:0}: Error finding container 094e230a91ee796a38f8b920ec9ab99bacf23906fbf2fa79ded8f87e45199fbc: Status 404 returned error can't find the container with id 094e230a91ee796a38f8b920ec9ab99bacf23906fbf2fa79ded8f87e45199fbc Oct 09 08:22:14 crc kubenswrapper[4872]: I1009 08:22:14.955258 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.121083 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x8nzf"] Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.691256 4872 generic.go:334] "Generic (PLEG): container finished" podID="f41d6540-7af0-4aa7-a36a-629e55c85a6f" containerID="319555a25121096bfc05f0a23c2ac6ee22146814049c160e9ac54254b2c6adbf" exitCode=0 Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.691372 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8nzf" event={"ID":"f41d6540-7af0-4aa7-a36a-629e55c85a6f","Type":"ContainerDied","Data":"319555a25121096bfc05f0a23c2ac6ee22146814049c160e9ac54254b2c6adbf"} Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.693097 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8nzf" event={"ID":"f41d6540-7af0-4aa7-a36a-629e55c85a6f","Type":"ContainerStarted","Data":"c7df3ad2061d271cff58aa7bf252eb00b2049b6cd922d3849d7f9fccbdfaf747"} Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.704870 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerID="8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1" exitCode=0 Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.705829 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerDied","Data":"8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1"} Oct 09 08:22:15 crc kubenswrapper[4872]: I1009 08:22:15.705876 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerStarted","Data":"094e230a91ee796a38f8b920ec9ab99bacf23906fbf2fa79ded8f87e45199fbc"} Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.714249 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8nzf" event={"ID":"f41d6540-7af0-4aa7-a36a-629e55c85a6f","Type":"ContainerStarted","Data":"e473cb37e2e21a3c40c7e335bb91292e0b106d3c3826cde76fdcb82a38bc8612"} Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.831836 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8swgj"] Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.832724 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.834592 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.842015 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8swgj"] Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.878968 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-utilities\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.879236 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-catalog-content\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.879498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4hd56\" (UniqueName: \"kubernetes.io/projected/67dc2250-a671-4f10-8ac9-46021dca556b-kube-api-access-4hd56\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.981088 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-utilities\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.980495 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-utilities\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.981288 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-catalog-content\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.981536 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4hd56\" (UniqueName: \"kubernetes.io/projected/67dc2250-a671-4f10-8ac9-46021dca556b-kube-api-access-4hd56\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:16 crc kubenswrapper[4872]: I1009 08:22:16.981738 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67dc2250-a671-4f10-8ac9-46021dca556b-catalog-content\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.003706 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4hd56\" (UniqueName: \"kubernetes.io/projected/67dc2250-a671-4f10-8ac9-46021dca556b-kube-api-access-4hd56\") pod \"redhat-marketplace-8swgj\" (UID: \"67dc2250-a671-4f10-8ac9-46021dca556b\") " pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.037035 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mdw5n"] Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.038331 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.041221 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.051161 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdw5n"] Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.083629 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj84n\" (UniqueName: \"kubernetes.io/projected/c9a7cca0-9eab-4de5-8100-98cf3f887af1-kube-api-access-zj84n\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.083711 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-catalog-content\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.083777 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-utilities\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.157682 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.185802 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-catalog-content\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.186014 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-utilities\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.186107 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj84n\" (UniqueName: \"kubernetes.io/projected/c9a7cca0-9eab-4de5-8100-98cf3f887af1-kube-api-access-zj84n\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.186336 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-catalog-content\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.186467 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9a7cca0-9eab-4de5-8100-98cf3f887af1-utilities\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.203062 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj84n\" (UniqueName: \"kubernetes.io/projected/c9a7cca0-9eab-4de5-8100-98cf3f887af1-kube-api-access-zj84n\") pod \"redhat-operators-mdw5n\" (UID: \"c9a7cca0-9eab-4de5-8100-98cf3f887af1\") " pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.363329 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.585505 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8swgj"] Oct 09 08:22:17 crc kubenswrapper[4872]: W1009 08:22:17.598928 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67dc2250_a671_4f10_8ac9_46021dca556b.slice/crio-4d993ebd9adf6816617d70d80c1873945e4701250a7a9fb37b460546a5da80c9 WatchSource:0}: Error finding container 4d993ebd9adf6816617d70d80c1873945e4701250a7a9fb37b460546a5da80c9: Status 404 returned error can't find the container with id 4d993ebd9adf6816617d70d80c1873945e4701250a7a9fb37b460546a5da80c9 Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.669173 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mdw5n"] Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.721982 4872 generic.go:334] "Generic (PLEG): container finished" podID="f41d6540-7af0-4aa7-a36a-629e55c85a6f" containerID="e473cb37e2e21a3c40c7e335bb91292e0b106d3c3826cde76fdcb82a38bc8612" exitCode=0 Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.724186 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8nzf" event={"ID":"f41d6540-7af0-4aa7-a36a-629e55c85a6f","Type":"ContainerDied","Data":"e473cb37e2e21a3c40c7e335bb91292e0b106d3c3826cde76fdcb82a38bc8612"} Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.725284 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8swgj" event={"ID":"67dc2250-a671-4f10-8ac9-46021dca556b","Type":"ContainerStarted","Data":"4d993ebd9adf6816617d70d80c1873945e4701250a7a9fb37b460546a5da80c9"} Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.727518 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerID="68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b" exitCode=0 Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.727597 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerDied","Data":"68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b"} Oct 09 08:22:17 crc kubenswrapper[4872]: I1009 08:22:17.728728 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdw5n" event={"ID":"c9a7cca0-9eab-4de5-8100-98cf3f887af1","Type":"ContainerStarted","Data":"312c07e430027036ccd4abf31b4fb6c101cad571f37828390a0ea70aaf9c4bdc"} Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.735341 4872 generic.go:334] "Generic (PLEG): container finished" podID="c9a7cca0-9eab-4de5-8100-98cf3f887af1" containerID="e922a223560d59d514aefcabba8d79e26112ee0737d9e438036bcde3e23257e7" exitCode=0 Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.735421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdw5n" event={"ID":"c9a7cca0-9eab-4de5-8100-98cf3f887af1","Type":"ContainerDied","Data":"e922a223560d59d514aefcabba8d79e26112ee0737d9e438036bcde3e23257e7"} Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.739224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x8nzf" event={"ID":"f41d6540-7af0-4aa7-a36a-629e55c85a6f","Type":"ContainerStarted","Data":"0c41f961be1c44688dbf234b60527e93757212281ae568ab1b7b5d396743b34c"} Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.740609 4872 generic.go:334] "Generic (PLEG): container finished" podID="67dc2250-a671-4f10-8ac9-46021dca556b" containerID="2b76cb535bc4332f960a58f27bae85dde12293628e8eb708e0a68862e51ea8ec" exitCode=0 Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.740714 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8swgj" event={"ID":"67dc2250-a671-4f10-8ac9-46021dca556b","Type":"ContainerDied","Data":"2b76cb535bc4332f960a58f27bae85dde12293628e8eb708e0a68862e51ea8ec"} Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.745546 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerStarted","Data":"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828"} Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.777217 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mxtxg" podStartSLOduration=2.209339564 podStartE2EDuration="4.777196243s" podCreationTimestamp="2025-10-09 08:22:14 +0000 UTC" firstStartedPulling="2025-10-09 08:22:15.712577173 +0000 UTC m=+293.903105789" lastFinishedPulling="2025-10-09 08:22:18.280433832 +0000 UTC m=+296.470962468" observedRunningTime="2025-10-09 08:22:18.776133451 +0000 UTC m=+296.966662067" watchObservedRunningTime="2025-10-09 08:22:18.777196243 +0000 UTC m=+296.967724869" Oct 09 08:22:18 crc kubenswrapper[4872]: I1009 08:22:18.813436 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x8nzf" podStartSLOduration=2.252553283 podStartE2EDuration="4.813414233s" podCreationTimestamp="2025-10-09 08:22:14 +0000 UTC" firstStartedPulling="2025-10-09 08:22:15.694447262 +0000 UTC m=+293.884975888" lastFinishedPulling="2025-10-09 08:22:18.255308212 +0000 UTC m=+296.445836838" observedRunningTime="2025-10-09 08:22:18.810656891 +0000 UTC m=+297.001185517" watchObservedRunningTime="2025-10-09 08:22:18.813414233 +0000 UTC m=+297.003942859" Oct 09 08:22:19 crc kubenswrapper[4872]: I1009 08:22:19.752354 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8swgj" event={"ID":"67dc2250-a671-4f10-8ac9-46021dca556b","Type":"ContainerStarted","Data":"1f8743a14afbc6efa4cc8b62261001c900bed153b12f7b10ee882f6d41462407"} Oct 09 08:22:19 crc kubenswrapper[4872]: I1009 08:22:19.754507 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdw5n" event={"ID":"c9a7cca0-9eab-4de5-8100-98cf3f887af1","Type":"ContainerStarted","Data":"e8085d14ff981121d32180f4b02e81035b18c642297f10e5690c3ceef6c7f07c"} Oct 09 08:22:20 crc kubenswrapper[4872]: I1009 08:22:20.760392 4872 generic.go:334] "Generic (PLEG): container finished" podID="c9a7cca0-9eab-4de5-8100-98cf3f887af1" containerID="e8085d14ff981121d32180f4b02e81035b18c642297f10e5690c3ceef6c7f07c" exitCode=0 Oct 09 08:22:20 crc kubenswrapper[4872]: I1009 08:22:20.760451 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdw5n" event={"ID":"c9a7cca0-9eab-4de5-8100-98cf3f887af1","Type":"ContainerDied","Data":"e8085d14ff981121d32180f4b02e81035b18c642297f10e5690c3ceef6c7f07c"} Oct 09 08:22:20 crc kubenswrapper[4872]: I1009 08:22:20.763892 4872 generic.go:334] "Generic (PLEG): container finished" podID="67dc2250-a671-4f10-8ac9-46021dca556b" containerID="1f8743a14afbc6efa4cc8b62261001c900bed153b12f7b10ee882f6d41462407" exitCode=0 Oct 09 08:22:20 crc kubenswrapper[4872]: I1009 08:22:20.763926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8swgj" event={"ID":"67dc2250-a671-4f10-8ac9-46021dca556b","Type":"ContainerDied","Data":"1f8743a14afbc6efa4cc8b62261001c900bed153b12f7b10ee882f6d41462407"} Oct 09 08:22:22 crc kubenswrapper[4872]: I1009 08:22:22.776536 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mdw5n" event={"ID":"c9a7cca0-9eab-4de5-8100-98cf3f887af1","Type":"ContainerStarted","Data":"b5d3f0949dafa56d99edb3648c1dc9661b75d276c4f129e52275280fb45da14d"} Oct 09 08:22:22 crc kubenswrapper[4872]: I1009 08:22:22.779706 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8swgj" event={"ID":"67dc2250-a671-4f10-8ac9-46021dca556b","Type":"ContainerStarted","Data":"9cf2525942f01bd2de67ba8fe3dd6132d556f61c633fa63f2312207e47e92a1a"} Oct 09 08:22:22 crc kubenswrapper[4872]: I1009 08:22:22.797891 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mdw5n" podStartSLOduration=2.929873642 podStartE2EDuration="5.797874576s" podCreationTimestamp="2025-10-09 08:22:17 +0000 UTC" firstStartedPulling="2025-10-09 08:22:18.73823171 +0000 UTC m=+296.928760336" lastFinishedPulling="2025-10-09 08:22:21.606232644 +0000 UTC m=+299.796761270" observedRunningTime="2025-10-09 08:22:22.794275388 +0000 UTC m=+300.984804004" watchObservedRunningTime="2025-10-09 08:22:22.797874576 +0000 UTC m=+300.988403202" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.752272 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.753108 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.798981 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.815912 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8swgj" podStartSLOduration=5.530312861 podStartE2EDuration="8.815890162s" podCreationTimestamp="2025-10-09 08:22:16 +0000 UTC" firstStartedPulling="2025-10-09 08:22:18.742400005 +0000 UTC m=+296.932928631" lastFinishedPulling="2025-10-09 08:22:22.027977306 +0000 UTC m=+300.218505932" observedRunningTime="2025-10-09 08:22:22.818058008 +0000 UTC m=+301.008586634" watchObservedRunningTime="2025-10-09 08:22:24.815890162 +0000 UTC m=+303.006418788" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.836166 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.956258 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.956657 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:24 crc kubenswrapper[4872]: I1009 08:22:24.992153 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:25 crc kubenswrapper[4872]: I1009 08:22:25.848311 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x8nzf" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.158553 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.158981 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.205437 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.363948 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.364009 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.415700 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.896344 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8swgj" Oct 09 08:22:27 crc kubenswrapper[4872]: I1009 08:22:27.899022 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mdw5n" Oct 09 08:23:08 crc kubenswrapper[4872]: I1009 08:23:08.157899 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:23:08 crc kubenswrapper[4872]: I1009 08:23:08.158421 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:23:38 crc kubenswrapper[4872]: I1009 08:23:38.157099 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:23:38 crc kubenswrapper[4872]: I1009 08:23:38.157967 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.157813 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.158350 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.158399 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.159077 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.159144 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f" gracePeriod=600 Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.349978 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f" exitCode=0 Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.350089 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f"} Oct 09 08:24:08 crc kubenswrapper[4872]: I1009 08:24:08.350562 4872 scope.go:117] "RemoveContainer" containerID="3ad7fd68e9f14a0d5aa4a6049525d4e501413f1687053b2784da61525145ae85" Oct 09 08:24:09 crc kubenswrapper[4872]: I1009 08:24:09.358918 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c"} Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.576849 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cnf9n"] Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.578205 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.598575 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cnf9n"] Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728138 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-registry-certificates\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728209 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709268b6-09b4-4578-9a32-de7865981bd9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728241 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709268b6-09b4-4578-9a32-de7865981bd9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-bound-sa-token\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728343 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9q48\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-kube-api-access-r9q48\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728376 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728406 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-registry-tls\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.728426 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-trusted-ca\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.753778 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829400 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709268b6-09b4-4578-9a32-de7865981bd9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-bound-sa-token\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9q48\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-kube-api-access-r9q48\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829544 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-registry-tls\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829564 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-trusted-ca\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829610 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-registry-certificates\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.829705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709268b6-09b4-4578-9a32-de7865981bd9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.830355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/709268b6-09b4-4578-9a32-de7865981bd9-ca-trust-extracted\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.831584 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-registry-certificates\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.832355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/709268b6-09b4-4578-9a32-de7865981bd9-trusted-ca\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.837518 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-registry-tls\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.838940 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/709268b6-09b4-4578-9a32-de7865981bd9-installation-pull-secrets\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.847379 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-bound-sa-token\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.850542 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9q48\" (UniqueName: \"kubernetes.io/projected/709268b6-09b4-4578-9a32-de7865981bd9-kube-api-access-r9q48\") pod \"image-registry-66df7c8f76-cnf9n\" (UID: \"709268b6-09b4-4578-9a32-de7865981bd9\") " pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:28 crc kubenswrapper[4872]: I1009 08:25:28.893822 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:29 crc kubenswrapper[4872]: I1009 08:25:29.280291 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-cnf9n"] Oct 09 08:25:29 crc kubenswrapper[4872]: I1009 08:25:29.832757 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" event={"ID":"709268b6-09b4-4578-9a32-de7865981bd9","Type":"ContainerStarted","Data":"42dababa0c8ac68bcaff4616cf0cbe918c8574af1dd9c0327718eb0e9586b80a"} Oct 09 08:25:29 crc kubenswrapper[4872]: I1009 08:25:29.832822 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" event={"ID":"709268b6-09b4-4578-9a32-de7865981bd9","Type":"ContainerStarted","Data":"4839996ee60b01d3c39fb20bc3d2b2b7bdd127189fac2243d2c602807c17eec5"} Oct 09 08:25:29 crc kubenswrapper[4872]: I1009 08:25:29.832941 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:29 crc kubenswrapper[4872]: I1009 08:25:29.854679 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" podStartSLOduration=1.854655233 podStartE2EDuration="1.854655233s" podCreationTimestamp="2025-10-09 08:25:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:25:29.851181711 +0000 UTC m=+488.041710347" watchObservedRunningTime="2025-10-09 08:25:29.854655233 +0000 UTC m=+488.045183859" Oct 09 08:25:48 crc kubenswrapper[4872]: I1009 08:25:48.900418 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-cnf9n" Oct 09 08:25:48 crc kubenswrapper[4872]: I1009 08:25:48.956158 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:26:08 crc kubenswrapper[4872]: I1009 08:26:08.156405 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:26:08 crc kubenswrapper[4872]: I1009 08:26:08.157025 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:26:13 crc kubenswrapper[4872]: I1009 08:26:13.991613 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" podUID="d734d590-eccc-42cc-ad8d-14e28f3c1882" containerName="registry" containerID="cri-o://70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60" gracePeriod=30 Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.319292 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442545 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442599 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442626 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grnsg\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442699 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442923 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.442968 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.443012 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.443063 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted\") pod \"d734d590-eccc-42cc-ad8d-14e28f3c1882\" (UID: \"d734d590-eccc-42cc-ad8d-14e28f3c1882\") " Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.444364 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.444433 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.449973 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.459212 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.460849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg" (OuterVolumeSpecName: "kube-api-access-grnsg") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "kube-api-access-grnsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.461015 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.461324 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.461335 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "d734d590-eccc-42cc-ad8d-14e28f3c1882" (UID: "d734d590-eccc-42cc-ad8d-14e28f3c1882"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.544978 4872 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/d734d590-eccc-42cc-ad8d-14e28f3c1882-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545051 4872 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545070 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grnsg\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-kube-api-access-grnsg\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545083 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545097 4872 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/d734d590-eccc-42cc-ad8d-14e28f3c1882-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545109 4872 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/d734d590-eccc-42cc-ad8d-14e28f3c1882-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:14 crc kubenswrapper[4872]: I1009 08:26:14.545120 4872 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/d734d590-eccc-42cc-ad8d-14e28f3c1882-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.101340 4872 generic.go:334] "Generic (PLEG): container finished" podID="d734d590-eccc-42cc-ad8d-14e28f3c1882" containerID="70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60" exitCode=0 Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.101387 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" event={"ID":"d734d590-eccc-42cc-ad8d-14e28f3c1882","Type":"ContainerDied","Data":"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60"} Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.101420 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" event={"ID":"d734d590-eccc-42cc-ad8d-14e28f3c1882","Type":"ContainerDied","Data":"c9f95989b6ade0b2ea2872e9f33e98eeaf9affb8c00bb0aaecc79925ccd09881"} Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.101436 4872 scope.go:117] "RemoveContainer" containerID="70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60" Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.101449 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-78fqw" Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.123266 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.127439 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-78fqw"] Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.131377 4872 scope.go:117] "RemoveContainer" containerID="70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60" Oct 09 08:26:15 crc kubenswrapper[4872]: E1009 08:26:15.132344 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60\": container with ID starting with 70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60 not found: ID does not exist" containerID="70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60" Oct 09 08:26:15 crc kubenswrapper[4872]: I1009 08:26:15.132373 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60"} err="failed to get container status \"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60\": rpc error: code = NotFound desc = could not find container \"70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60\": container with ID starting with 70c77d2cb4488d660d8249d9cda06e7e74cc33d6dc6c6a1d855d6affd1765c60 not found: ID does not exist" Oct 09 08:26:16 crc kubenswrapper[4872]: I1009 08:26:16.469032 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d734d590-eccc-42cc-ad8d-14e28f3c1882" path="/var/lib/kubelet/pods/d734d590-eccc-42cc-ad8d-14e28f3c1882/volumes" Oct 09 08:26:38 crc kubenswrapper[4872]: I1009 08:26:38.157809 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:26:38 crc kubenswrapper[4872]: I1009 08:26:38.158757 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.157308 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.157954 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.158003 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.158534 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.158584 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c" gracePeriod=600 Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.436849 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c" exitCode=0 Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.436903 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c"} Oct 09 08:27:08 crc kubenswrapper[4872]: I1009 08:27:08.437596 4872 scope.go:117] "RemoveContainer" containerID="fb09122f972ce857cba1d7421d0cc7c3774851f979874ac60b01dee3ba6e670f" Oct 09 08:27:09 crc kubenswrapper[4872]: I1009 08:27:09.448506 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd"} Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.557426 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r594q"] Oct 09 08:28:16 crc kubenswrapper[4872]: E1009 08:28:16.558172 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d734d590-eccc-42cc-ad8d-14e28f3c1882" containerName="registry" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.558185 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d734d590-eccc-42cc-ad8d-14e28f3c1882" containerName="registry" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.558280 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d734d590-eccc-42cc-ad8d-14e28f3c1882" containerName="registry" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.558688 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.562789 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.562824 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-jhhlx" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.562802 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.577622 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r594q"] Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.599725 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cq5wd"] Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.600384 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-cq5wd" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.603225 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-fqgmz" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.608054 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k456p"] Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.609763 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.616032 4872 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-8zfm6" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.629523 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cq5wd"] Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.641028 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k456p"] Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.679510 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cm72d\" (UniqueName: \"kubernetes.io/projected/eb652e7d-6150-46ae-956f-0fcf642d935a-kube-api-access-cm72d\") pod \"cert-manager-cainjector-7f985d654d-r594q\" (UID: \"eb652e7d-6150-46ae-956f-0fcf642d935a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.781062 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n8wdn\" (UniqueName: \"kubernetes.io/projected/f1f5f354-df1e-49fa-a117-8e99befeaf38-kube-api-access-n8wdn\") pod \"cert-manager-webhook-5655c58dd6-k456p\" (UID: \"f1f5f354-df1e-49fa-a117-8e99befeaf38\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.781604 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5pkt\" (UniqueName: \"kubernetes.io/projected/ba374ef8-6f28-4a47-a948-8cd66cff6132-kube-api-access-p5pkt\") pod \"cert-manager-5b446d88c5-cq5wd\" (UID: \"ba374ef8-6f28-4a47-a948-8cd66cff6132\") " pod="cert-manager/cert-manager-5b446d88c5-cq5wd" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.781935 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cm72d\" (UniqueName: \"kubernetes.io/projected/eb652e7d-6150-46ae-956f-0fcf642d935a-kube-api-access-cm72d\") pod \"cert-manager-cainjector-7f985d654d-r594q\" (UID: \"eb652e7d-6150-46ae-956f-0fcf642d935a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.800535 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cm72d\" (UniqueName: \"kubernetes.io/projected/eb652e7d-6150-46ae-956f-0fcf642d935a-kube-api-access-cm72d\") pod \"cert-manager-cainjector-7f985d654d-r594q\" (UID: \"eb652e7d-6150-46ae-956f-0fcf642d935a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.878928 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.883551 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n8wdn\" (UniqueName: \"kubernetes.io/projected/f1f5f354-df1e-49fa-a117-8e99befeaf38-kube-api-access-n8wdn\") pod \"cert-manager-webhook-5655c58dd6-k456p\" (UID: \"f1f5f354-df1e-49fa-a117-8e99befeaf38\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.883760 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5pkt\" (UniqueName: \"kubernetes.io/projected/ba374ef8-6f28-4a47-a948-8cd66cff6132-kube-api-access-p5pkt\") pod \"cert-manager-5b446d88c5-cq5wd\" (UID: \"ba374ef8-6f28-4a47-a948-8cd66cff6132\") " pod="cert-manager/cert-manager-5b446d88c5-cq5wd" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.900527 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n8wdn\" (UniqueName: \"kubernetes.io/projected/f1f5f354-df1e-49fa-a117-8e99befeaf38-kube-api-access-n8wdn\") pod \"cert-manager-webhook-5655c58dd6-k456p\" (UID: \"f1f5f354-df1e-49fa-a117-8e99befeaf38\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.901479 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5pkt\" (UniqueName: \"kubernetes.io/projected/ba374ef8-6f28-4a47-a948-8cd66cff6132-kube-api-access-p5pkt\") pod \"cert-manager-5b446d88c5-cq5wd\" (UID: \"ba374ef8-6f28-4a47-a948-8cd66cff6132\") " pod="cert-manager/cert-manager-5b446d88c5-cq5wd" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.921251 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-cq5wd" Oct 09 08:28:16 crc kubenswrapper[4872]: I1009 08:28:16.931749 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.101914 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r594q"] Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.106989 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.178331 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k456p"] Oct 09 08:28:17 crc kubenswrapper[4872]: W1009 08:28:17.186389 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1f5f354_df1e_49fa_a117_8e99befeaf38.slice/crio-8686d1489f92133523e76dfe14686c82140d13ec4b188d8a85fef0abb0036ec8 WatchSource:0}: Error finding container 8686d1489f92133523e76dfe14686c82140d13ec4b188d8a85fef0abb0036ec8: Status 404 returned error can't find the container with id 8686d1489f92133523e76dfe14686c82140d13ec4b188d8a85fef0abb0036ec8 Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.414048 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-cq5wd"] Oct 09 08:28:17 crc kubenswrapper[4872]: W1009 08:28:17.426307 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba374ef8_6f28_4a47_a948_8cd66cff6132.slice/crio-51d19f5d9d1ff073746bc02da93dcbe75b0b91e23245c11179069ba851d11786 WatchSource:0}: Error finding container 51d19f5d9d1ff073746bc02da93dcbe75b0b91e23245c11179069ba851d11786: Status 404 returned error can't find the container with id 51d19f5d9d1ff073746bc02da93dcbe75b0b91e23245c11179069ba851d11786 Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.869113 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-cq5wd" event={"ID":"ba374ef8-6f28-4a47-a948-8cd66cff6132","Type":"ContainerStarted","Data":"51d19f5d9d1ff073746bc02da93dcbe75b0b91e23245c11179069ba851d11786"} Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.870584 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" event={"ID":"eb652e7d-6150-46ae-956f-0fcf642d935a","Type":"ContainerStarted","Data":"ec0fe050c1e91b7fc8577cd21afdc51039e6ad8c8beb17c88cf96139a5df8e48"} Oct 09 08:28:17 crc kubenswrapper[4872]: I1009 08:28:17.871991 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" event={"ID":"f1f5f354-df1e-49fa-a117-8e99befeaf38","Type":"ContainerStarted","Data":"8686d1489f92133523e76dfe14686c82140d13ec4b188d8a85fef0abb0036ec8"} Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.892871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-cq5wd" event={"ID":"ba374ef8-6f28-4a47-a948-8cd66cff6132","Type":"ContainerStarted","Data":"6d59853f892f093a4759ddd90bb7950ae117dc077fce7e1c58af7c46ee344a82"} Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.896022 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" event={"ID":"eb652e7d-6150-46ae-956f-0fcf642d935a","Type":"ContainerStarted","Data":"10393caaa98fdf6c837328b88e576fa377a8b2df832b156f124acbc8823dc3b7"} Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.897423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" event={"ID":"f1f5f354-df1e-49fa-a117-8e99befeaf38","Type":"ContainerStarted","Data":"d465beb1118205c905c28226faba54b51ebc6e46bd210a4289ba17f86976a0db"} Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.897756 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.923005 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-cq5wd" podStartSLOduration=1.764377698 podStartE2EDuration="4.922986237s" podCreationTimestamp="2025-10-09 08:28:16 +0000 UTC" firstStartedPulling="2025-10-09 08:28:17.428532843 +0000 UTC m=+655.619061469" lastFinishedPulling="2025-10-09 08:28:20.587141382 +0000 UTC m=+658.777670008" observedRunningTime="2025-10-09 08:28:20.908205969 +0000 UTC m=+659.098734595" watchObservedRunningTime="2025-10-09 08:28:20.922986237 +0000 UTC m=+659.113514863" Oct 09 08:28:20 crc kubenswrapper[4872]: I1009 08:28:20.940563 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" podStartSLOduration=2.203709088 podStartE2EDuration="4.940527855s" podCreationTimestamp="2025-10-09 08:28:16 +0000 UTC" firstStartedPulling="2025-10-09 08:28:17.190944453 +0000 UTC m=+655.381473079" lastFinishedPulling="2025-10-09 08:28:19.92776322 +0000 UTC m=+658.118291846" observedRunningTime="2025-10-09 08:28:20.926402486 +0000 UTC m=+659.116931122" watchObservedRunningTime="2025-10-09 08:28:20.940527855 +0000 UTC m=+659.131056491" Oct 09 08:28:26 crc kubenswrapper[4872]: I1009 08:28:26.935616 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-k456p" Oct 09 08:28:26 crc kubenswrapper[4872]: I1009 08:28:26.957282 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-r594q" podStartSLOduration=7.536272555 podStartE2EDuration="10.957263743s" podCreationTimestamp="2025-10-09 08:28:16 +0000 UTC" firstStartedPulling="2025-10-09 08:28:17.106617081 +0000 UTC m=+655.297145707" lastFinishedPulling="2025-10-09 08:28:20.527608269 +0000 UTC m=+658.718136895" observedRunningTime="2025-10-09 08:28:20.937822257 +0000 UTC m=+659.128350883" watchObservedRunningTime="2025-10-09 08:28:26.957263743 +0000 UTC m=+665.147792369" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.195262 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xnhdk"] Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196255 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-controller" containerID="cri-o://b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196278 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="nbdb" containerID="cri-o://879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196311 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="northd" containerID="cri-o://188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196402 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-acl-logging" containerID="cri-o://1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196383 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-node" containerID="cri-o://95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196489 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="sbdb" containerID="cri-o://733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.196580 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.230206 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" containerID="cri-o://1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" gracePeriod=30 Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.906744 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/3.log" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.910383 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovn-acl-logging/0.log" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.911347 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovn-controller/0.log" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.912364 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.987897 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nbjl5"] Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988141 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988155 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988168 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988176 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988188 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="northd" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988196 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="northd" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988211 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kubecfg-setup" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988219 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kubecfg-setup" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988232 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="sbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988241 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="sbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988256 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988264 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988283 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-acl-logging" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988291 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-acl-logging" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988304 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-node" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988313 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-node" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988326 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988334 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988345 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988352 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988365 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="nbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988373 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="nbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988500 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988511 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-acl-logging" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988525 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="nbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988534 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="sbdb" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988544 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-node" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988555 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovn-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988564 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="kube-rbac-proxy-ovn-metrics" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988577 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="northd" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988590 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988600 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.988745 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988756 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988872 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.988888 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: E1009 08:28:43.989010 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.989019 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" containerName="ovnkube-controller" Oct 09 08:28:43 crc kubenswrapper[4872]: I1009 08:28:43.991444 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.038768 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/2.log" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.039650 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/1.log" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.039709 4872 generic.go:334] "Generic (PLEG): container finished" podID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" containerID="6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd" exitCode=2 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.039779 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerDied","Data":"6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.039817 4872 scope.go:117] "RemoveContainer" containerID="7a2662e49d2b52034ff6540dc206aeb5a5ff153010fd1c44b5e1a1b6319c4cbc" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.040728 4872 scope.go:117] "RemoveContainer" containerID="6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.041048 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5ftzb_openshift-multus(d5f349b9-7850-4b1b-86da-e89f7cfd32df)\"" pod="openshift-multus/multus-5ftzb" podUID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048299 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovnkube-controller/3.log" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048430 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048114 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048687 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log" (OuterVolumeSpecName: "node-log") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048705 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048778 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048813 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048885 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048908 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048982 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-85kdn\" (UniqueName: \"kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049128 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049152 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049180 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049204 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049229 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049255 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049284 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049320 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049404 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049451 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049473 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049501 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049522 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049542 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch\") pod \"fac60b03-54bb-43e8-8994-3674d42b209b\" (UID: \"fac60b03-54bb-43e8-8994-3674d42b209b\") " Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050060 4872 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050076 4872 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-node-log\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050088 4872 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.048922 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.049587 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050161 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050187 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050178 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050209 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050232 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050262 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket" (OuterVolumeSpecName: "log-socket") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050287 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.050316 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.051333 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.051793 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.051841 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash" (OuterVolumeSpecName: "host-slash") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.052351 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.056965 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovn-acl-logging/0.log" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.057790 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-xnhdk_fac60b03-54bb-43e8-8994-3674d42b209b/ovn-controller/0.log" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.060807 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.064985 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065027 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065036 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065044 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065054 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065060 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" exitCode=0 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065067 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" exitCode=143 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065075 4872 generic.go:334] "Generic (PLEG): container finished" podID="fac60b03-54bb-43e8-8994-3674d42b209b" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" exitCode=143 Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065101 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065135 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065166 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065178 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065194 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065209 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065223 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065220 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065232 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065386 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065394 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065399 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065405 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065413 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065420 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065427 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065439 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065452 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065462 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065470 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065477 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065485 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065493 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065501 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065508 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065516 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.065524 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066579 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066605 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066616 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066622 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066628 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066651 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066656 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066663 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066669 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066674 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066680 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066689 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-xnhdk" event={"ID":"fac60b03-54bb-43e8-8994-3674d42b209b","Type":"ContainerDied","Data":"1e57250e14b035a83086e2213c2227f1cf0afb4cbf994c4184bbaad1bdf1f7be"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066698 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066704 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066709 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066715 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066720 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066725 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066731 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066736 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066742 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.066748 4872 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.067075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn" (OuterVolumeSpecName: "kube-api-access-85kdn") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "kube-api-access-85kdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.073203 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "fac60b03-54bb-43e8-8994-3674d42b209b" (UID: "fac60b03-54bb-43e8-8994-3674d42b209b"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.074042 4872 scope.go:117] "RemoveContainer" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.095367 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.121905 4872 scope.go:117] "RemoveContainer" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.139615 4872 scope.go:117] "RemoveContainer" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.151708 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-systemd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.151883 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-kubelet\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.151986 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-netd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152073 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-slash\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152185 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-var-lib-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152291 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84791e42-6e2d-4696-b515-36690b0409fa-ovn-node-metrics-cert\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152367 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-node-log\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152447 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-env-overrides\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152528 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-ovn\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152625 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-config\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152788 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-netns\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.152883 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-script-lib\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153012 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8gwl\" (UniqueName: \"kubernetes.io/projected/84791e42-6e2d-4696-b515-36690b0409fa-kube-api-access-f8gwl\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153118 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-bin\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153204 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153287 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-systemd-units\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-log-socket\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153538 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153547 4872 scope.go:117] "RemoveContainer" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153651 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153790 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-etc-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153864 4872 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153881 4872 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153893 4872 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153905 4872 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153917 4872 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153927 4872 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153939 4872 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153951 4872 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153962 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-85kdn\" (UniqueName: \"kubernetes.io/projected/fac60b03-54bb-43e8-8994-3674d42b209b-kube-api-access-85kdn\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153973 4872 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153985 4872 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.153997 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.154009 4872 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-host-slash\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.154020 4872 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.154029 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/fac60b03-54bb-43e8-8994-3674d42b209b-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.154039 4872 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/fac60b03-54bb-43e8-8994-3674d42b209b-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.154048 4872 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/fac60b03-54bb-43e8-8994-3674d42b209b-log-socket\") on node \"crc\" DevicePath \"\"" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.172952 4872 scope.go:117] "RemoveContainer" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.200170 4872 scope.go:117] "RemoveContainer" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.214579 4872 scope.go:117] "RemoveContainer" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.229095 4872 scope.go:117] "RemoveContainer" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.242707 4872 scope.go:117] "RemoveContainer" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254769 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-systemd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254833 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-kubelet\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254841 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-systemd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254865 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-netd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254896 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-netd\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254903 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-slash\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.254933 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-slash\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-var-lib-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255036 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-kubelet\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255079 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-var-lib-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255065 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84791e42-6e2d-4696-b515-36690b0409fa-ovn-node-metrics-cert\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255210 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-node-log\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255243 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-env-overrides\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255269 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-ovn\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255307 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-config\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255322 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-node-log\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-netns\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255386 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-script-lib\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255439 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-ovn\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255441 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-netns\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255566 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8gwl\" (UniqueName: \"kubernetes.io/projected/84791e42-6e2d-4696-b515-36690b0409fa-kube-api-access-f8gwl\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255614 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255632 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-bin\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255709 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-log-socket\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255733 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-systemd-units\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255759 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255801 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255823 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-etc-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255910 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-etc-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255940 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-run-openvswitch\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255979 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-run-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.255987 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-systemd-units\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256002 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256027 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-log-socket\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256031 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/84791e42-6e2d-4696-b515-36690b0409fa-host-cni-bin\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256079 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-env-overrides\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256219 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-config\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.256455 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/84791e42-6e2d-4696-b515-36690b0409fa-ovnkube-script-lib\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.258173 4872 scope.go:117] "RemoveContainer" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.258749 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/84791e42-6e2d-4696-b515-36690b0409fa-ovn-node-metrics-cert\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.259147 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": container with ID starting with 1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5 not found: ID does not exist" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.259189 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} err="failed to get container status \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": rpc error: code = NotFound desc = could not find container \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": container with ID starting with 1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.259225 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.259831 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": container with ID starting with 7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d not found: ID does not exist" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.259952 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} err="failed to get container status \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": rpc error: code = NotFound desc = could not find container \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": container with ID starting with 7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.259973 4872 scope.go:117] "RemoveContainer" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.260222 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": container with ID starting with 733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c not found: ID does not exist" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.260254 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} err="failed to get container status \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": rpc error: code = NotFound desc = could not find container \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": container with ID starting with 733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.260272 4872 scope.go:117] "RemoveContainer" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.260616 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": container with ID starting with 879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8 not found: ID does not exist" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.260901 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} err="failed to get container status \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": rpc error: code = NotFound desc = could not find container \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": container with ID starting with 879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.260927 4872 scope.go:117] "RemoveContainer" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.261233 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": container with ID starting with 188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5 not found: ID does not exist" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.261395 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} err="failed to get container status \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": rpc error: code = NotFound desc = could not find container \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": container with ID starting with 188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.261510 4872 scope.go:117] "RemoveContainer" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.262214 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": container with ID starting with ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6 not found: ID does not exist" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.262254 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} err="failed to get container status \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": rpc error: code = NotFound desc = could not find container \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": container with ID starting with ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.262274 4872 scope.go:117] "RemoveContainer" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.262584 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": container with ID starting with 95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7 not found: ID does not exist" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.262728 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} err="failed to get container status \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": rpc error: code = NotFound desc = could not find container \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": container with ID starting with 95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.262830 4872 scope.go:117] "RemoveContainer" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.263203 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": container with ID starting with 1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e not found: ID does not exist" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.263235 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} err="failed to get container status \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": rpc error: code = NotFound desc = could not find container \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": container with ID starting with 1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.263255 4872 scope.go:117] "RemoveContainer" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.263519 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": container with ID starting with b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad not found: ID does not exist" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.263629 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} err="failed to get container status \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": rpc error: code = NotFound desc = could not find container \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": container with ID starting with b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.263750 4872 scope.go:117] "RemoveContainer" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: E1009 08:28:44.264129 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": container with ID starting with 0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f not found: ID does not exist" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264168 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} err="failed to get container status \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": rpc error: code = NotFound desc = could not find container \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": container with ID starting with 0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264191 4872 scope.go:117] "RemoveContainer" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264454 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} err="failed to get container status \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": rpc error: code = NotFound desc = could not find container \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": container with ID starting with 1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264538 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264953 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} err="failed to get container status \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": rpc error: code = NotFound desc = could not find container \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": container with ID starting with 7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.264982 4872 scope.go:117] "RemoveContainer" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265247 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} err="failed to get container status \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": rpc error: code = NotFound desc = could not find container \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": container with ID starting with 733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265295 4872 scope.go:117] "RemoveContainer" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265570 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} err="failed to get container status \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": rpc error: code = NotFound desc = could not find container \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": container with ID starting with 879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265592 4872 scope.go:117] "RemoveContainer" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265893 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} err="failed to get container status \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": rpc error: code = NotFound desc = could not find container \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": container with ID starting with 188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.265923 4872 scope.go:117] "RemoveContainer" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.266221 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} err="failed to get container status \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": rpc error: code = NotFound desc = could not find container \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": container with ID starting with ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.266250 4872 scope.go:117] "RemoveContainer" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.266535 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} err="failed to get container status \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": rpc error: code = NotFound desc = could not find container \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": container with ID starting with 95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.266632 4872 scope.go:117] "RemoveContainer" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.267064 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} err="failed to get container status \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": rpc error: code = NotFound desc = could not find container \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": container with ID starting with 1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.267259 4872 scope.go:117] "RemoveContainer" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.267651 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} err="failed to get container status \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": rpc error: code = NotFound desc = could not find container \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": container with ID starting with b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.267689 4872 scope.go:117] "RemoveContainer" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268081 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} err="failed to get container status \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": rpc error: code = NotFound desc = could not find container \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": container with ID starting with 0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268124 4872 scope.go:117] "RemoveContainer" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268513 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} err="failed to get container status \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": rpc error: code = NotFound desc = could not find container \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": container with ID starting with 1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268544 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268873 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} err="failed to get container status \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": rpc error: code = NotFound desc = could not find container \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": container with ID starting with 7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.268897 4872 scope.go:117] "RemoveContainer" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.269170 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} err="failed to get container status \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": rpc error: code = NotFound desc = could not find container \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": container with ID starting with 733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.269193 4872 scope.go:117] "RemoveContainer" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.270743 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} err="failed to get container status \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": rpc error: code = NotFound desc = could not find container \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": container with ID starting with 879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.270766 4872 scope.go:117] "RemoveContainer" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271044 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} err="failed to get container status \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": rpc error: code = NotFound desc = could not find container \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": container with ID starting with 188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271069 4872 scope.go:117] "RemoveContainer" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271446 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} err="failed to get container status \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": rpc error: code = NotFound desc = could not find container \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": container with ID starting with ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271465 4872 scope.go:117] "RemoveContainer" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271734 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} err="failed to get container status \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": rpc error: code = NotFound desc = could not find container \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": container with ID starting with 95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271755 4872 scope.go:117] "RemoveContainer" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.271976 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} err="failed to get container status \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": rpc error: code = NotFound desc = could not find container \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": container with ID starting with 1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272007 4872 scope.go:117] "RemoveContainer" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272275 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} err="failed to get container status \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": rpc error: code = NotFound desc = could not find container \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": container with ID starting with b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272300 4872 scope.go:117] "RemoveContainer" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272517 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} err="failed to get container status \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": rpc error: code = NotFound desc = could not find container \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": container with ID starting with 0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272552 4872 scope.go:117] "RemoveContainer" containerID="1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272955 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5"} err="failed to get container status \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": rpc error: code = NotFound desc = could not find container \"1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5\": container with ID starting with 1927824f5be350d7b92e4b55a1ea8370962e4f0c84d789de04807546af70b9f5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.272986 4872 scope.go:117] "RemoveContainer" containerID="7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273249 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d"} err="failed to get container status \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": rpc error: code = NotFound desc = could not find container \"7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d\": container with ID starting with 7dc3aa65b5f810c4f46174890035cc64e8ce9b321e4f1fa64f45e95874d6475d not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273300 4872 scope.go:117] "RemoveContainer" containerID="733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273545 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c"} err="failed to get container status \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": rpc error: code = NotFound desc = could not find container \"733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c\": container with ID starting with 733973fb447e372242068cd45d7ef321b00dd38bf805117bb5f70ccca6698b0c not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273574 4872 scope.go:117] "RemoveContainer" containerID="879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273914 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8"} err="failed to get container status \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": rpc error: code = NotFound desc = could not find container \"879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8\": container with ID starting with 879a74c199ed21468655f22dec4aeb74792b7aa86c9b3323b752bee34b7105d8 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.273938 4872 scope.go:117] "RemoveContainer" containerID="188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274258 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5"} err="failed to get container status \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": rpc error: code = NotFound desc = could not find container \"188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5\": container with ID starting with 188d7f665a30157f79c2342e05d7219ab349c945d275f20b30f3ab61be616ae5 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274287 4872 scope.go:117] "RemoveContainer" containerID="ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274549 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6"} err="failed to get container status \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": rpc error: code = NotFound desc = could not find container \"ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6\": container with ID starting with ab5de16de4deaf52256679b73bc44beb9d1c7f355d8470d5a5e211b826516ba6 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274574 4872 scope.go:117] "RemoveContainer" containerID="95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274824 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7"} err="failed to get container status \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": rpc error: code = NotFound desc = could not find container \"95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7\": container with ID starting with 95f74bd14e3dbb7d04376ca49f90e8d0baa302ec2ed8660794db88ca41aec9b7 not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.274849 4872 scope.go:117] "RemoveContainer" containerID="1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275087 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e"} err="failed to get container status \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": rpc error: code = NotFound desc = could not find container \"1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e\": container with ID starting with 1a5cbc07878ddd662f443ab22cc56cf7aaa58391eb5f5958bcd866e838c1500e not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275116 4872 scope.go:117] "RemoveContainer" containerID="b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275251 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8gwl\" (UniqueName: \"kubernetes.io/projected/84791e42-6e2d-4696-b515-36690b0409fa-kube-api-access-f8gwl\") pod \"ovnkube-node-nbjl5\" (UID: \"84791e42-6e2d-4696-b515-36690b0409fa\") " pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275429 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad"} err="failed to get container status \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": rpc error: code = NotFound desc = could not find container \"b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad\": container with ID starting with b81c0e4846161691c228a28ea71a307fc00e6dc624cf9e9a98d2d9b54a14bfad not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275450 4872 scope.go:117] "RemoveContainer" containerID="0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.275744 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f"} err="failed to get container status \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": rpc error: code = NotFound desc = could not find container \"0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f\": container with ID starting with 0ae73450ba40dd5cef367d8aebbf296126f70eda133b2fd90b490e67cc094b2f not found: ID does not exist" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.308502 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.417394 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xnhdk"] Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.421000 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-xnhdk"] Oct 09 08:28:44 crc kubenswrapper[4872]: I1009 08:28:44.472546 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fac60b03-54bb-43e8-8994-3674d42b209b" path="/var/lib/kubelet/pods/fac60b03-54bb-43e8-8994-3674d42b209b/volumes" Oct 09 08:28:45 crc kubenswrapper[4872]: I1009 08:28:45.092917 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/2.log" Oct 09 08:28:45 crc kubenswrapper[4872]: I1009 08:28:45.095929 4872 generic.go:334] "Generic (PLEG): container finished" podID="84791e42-6e2d-4696-b515-36690b0409fa" containerID="da6471f0a76a226265677fa29b4c623720a6e7ff10e43701faddcfd5fd2dbdfc" exitCode=0 Oct 09 08:28:45 crc kubenswrapper[4872]: I1009 08:28:45.096020 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerDied","Data":"da6471f0a76a226265677fa29b4c623720a6e7ff10e43701faddcfd5fd2dbdfc"} Oct 09 08:28:45 crc kubenswrapper[4872]: I1009 08:28:45.096059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"303f2a4bc318c3117a25e36e1f4a05ca8a3ad99d27363e6028a99635c8933b7f"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110105 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"a0037c0ad3cee174a0d32a153230ea81bb84e3f432a4041d72d9d3743edf8bc7"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110456 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"8a7a372b5a0dff48a1e478b28d88375900e2c6ab0f0ec07ed046a8c4ec684399"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110470 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"db3241088af026a7534ebdd2e478597e0c9c81d35458947e663bf4b041223db1"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110479 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"b40a1aed3a43997688c335ca8edd2d11ff28b6688a9677509b8d353f6683de7a"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110487 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"6bddd977e4a572afe24ab95144db87680e6ebcecbe6b07348b1a7ab295b4f75f"} Oct 09 08:28:46 crc kubenswrapper[4872]: I1009 08:28:46.110495 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"4540091806d4dae38444a336ee38683ce8d37b8faf29ca7177bb7660a0f4f572"} Oct 09 08:28:48 crc kubenswrapper[4872]: I1009 08:28:48.126685 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"78bfdf8f0532642f0b969650a9bd50f02518f80dd824169649b0beeb064b0d52"} Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.148824 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" event={"ID":"84791e42-6e2d-4696-b515-36690b0409fa","Type":"ContainerStarted","Data":"244f9b7b261793f4fc2d1d3c75c9cbbfe1c984276d1faedfd76a0ee379e59048"} Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.149216 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.149279 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.149295 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.174275 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.175271 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:28:51 crc kubenswrapper[4872]: I1009 08:28:51.179403 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" podStartSLOduration=8.17938399 podStartE2EDuration="8.17938399s" podCreationTimestamp="2025-10-09 08:28:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:28:51.175765745 +0000 UTC m=+689.366294401" watchObservedRunningTime="2025-10-09 08:28:51.17938399 +0000 UTC m=+689.369912626" Oct 09 08:28:59 crc kubenswrapper[4872]: I1009 08:28:59.461560 4872 scope.go:117] "RemoveContainer" containerID="6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd" Oct 09 08:28:59 crc kubenswrapper[4872]: E1009 08:28:59.465580 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-5ftzb_openshift-multus(d5f349b9-7850-4b1b-86da-e89f7cfd32df)\"" pod="openshift-multus/multus-5ftzb" podUID="d5f349b9-7850-4b1b-86da-e89f7cfd32df" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.159752 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn"] Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.161094 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.162876 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.170401 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn"] Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.265845 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tbk5\" (UniqueName: \"kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.266271 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.266300 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.367907 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.367957 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.368021 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tbk5\" (UniqueName: \"kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.368481 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.368572 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.400993 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tbk5\" (UniqueName: \"kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: I1009 08:29:07.539050 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: E1009 08:29:07.572042 4872 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(659f684f7657aa972c4f0b623f1f22e47a2926f54fca9d775e8af36985ea211c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 08:29:07 crc kubenswrapper[4872]: E1009 08:29:07.572119 4872 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(659f684f7657aa972c4f0b623f1f22e47a2926f54fca9d775e8af36985ea211c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: E1009 08:29:07.572145 4872 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(659f684f7657aa972c4f0b623f1f22e47a2926f54fca9d775e8af36985ea211c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:07 crc kubenswrapper[4872]: E1009 08:29:07.572196 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace(d492de09-5100-4997-b375-2e60812f4591)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace(d492de09-5100-4997-b375-2e60812f4591)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(659f684f7657aa972c4f0b623f1f22e47a2926f54fca9d775e8af36985ea211c): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" podUID="d492de09-5100-4997-b375-2e60812f4591" Oct 09 08:29:08 crc kubenswrapper[4872]: I1009 08:29:08.157196 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:29:08 crc kubenswrapper[4872]: I1009 08:29:08.157290 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:29:08 crc kubenswrapper[4872]: I1009 08:29:08.246794 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:08 crc kubenswrapper[4872]: I1009 08:29:08.247362 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:08 crc kubenswrapper[4872]: E1009 08:29:08.271973 4872 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(e59cd2b201d2aa741ab4e31d188d744413603d97be0af6938a3dc25b5785a79b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 09 08:29:08 crc kubenswrapper[4872]: E1009 08:29:08.272105 4872 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(e59cd2b201d2aa741ab4e31d188d744413603d97be0af6938a3dc25b5785a79b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:08 crc kubenswrapper[4872]: E1009 08:29:08.272150 4872 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(e59cd2b201d2aa741ab4e31d188d744413603d97be0af6938a3dc25b5785a79b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:08 crc kubenswrapper[4872]: E1009 08:29:08.272239 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace(d492de09-5100-4997-b375-2e60812f4591)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace(d492de09-5100-4997-b375-2e60812f4591)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_openshift-marketplace_d492de09-5100-4997-b375-2e60812f4591_0(e59cd2b201d2aa741ab4e31d188d744413603d97be0af6938a3dc25b5785a79b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" podUID="d492de09-5100-4997-b375-2e60812f4591" Oct 09 08:29:13 crc kubenswrapper[4872]: I1009 08:29:13.462469 4872 scope.go:117] "RemoveContainer" containerID="6c4649e4f513debc659aaf1cf235dfa1294cd1f958902ef2f3e0d38a39b47bbd" Oct 09 08:29:14 crc kubenswrapper[4872]: I1009 08:29:14.285259 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5ftzb_d5f349b9-7850-4b1b-86da-e89f7cfd32df/kube-multus/2.log" Oct 09 08:29:14 crc kubenswrapper[4872]: I1009 08:29:14.285688 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5ftzb" event={"ID":"d5f349b9-7850-4b1b-86da-e89f7cfd32df","Type":"ContainerStarted","Data":"5895148ee05cbbcb32272b50b99fc2176cb8b748f5692f323ce51e935a06f23c"} Oct 09 08:29:14 crc kubenswrapper[4872]: I1009 08:29:14.390519 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nbjl5" Oct 09 08:29:23 crc kubenswrapper[4872]: I1009 08:29:23.460949 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:23 crc kubenswrapper[4872]: I1009 08:29:23.462350 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:23 crc kubenswrapper[4872]: I1009 08:29:23.704685 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn"] Oct 09 08:29:23 crc kubenswrapper[4872]: W1009 08:29:23.720867 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd492de09_5100_4997_b375_2e60812f4591.slice/crio-ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e WatchSource:0}: Error finding container ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e: Status 404 returned error can't find the container with id ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e Oct 09 08:29:24 crc kubenswrapper[4872]: I1009 08:29:24.350162 4872 generic.go:334] "Generic (PLEG): container finished" podID="d492de09-5100-4997-b375-2e60812f4591" containerID="e14236e3db0f37039e386534079071ea6040a5e01dfcd2f7abb8399d0e1fd9e1" exitCode=0 Oct 09 08:29:24 crc kubenswrapper[4872]: I1009 08:29:24.350234 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" event={"ID":"d492de09-5100-4997-b375-2e60812f4591","Type":"ContainerDied","Data":"e14236e3db0f37039e386534079071ea6040a5e01dfcd2f7abb8399d0e1fd9e1"} Oct 09 08:29:24 crc kubenswrapper[4872]: I1009 08:29:24.350546 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" event={"ID":"d492de09-5100-4997-b375-2e60812f4591","Type":"ContainerStarted","Data":"ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e"} Oct 09 08:29:26 crc kubenswrapper[4872]: I1009 08:29:26.365251 4872 generic.go:334] "Generic (PLEG): container finished" podID="d492de09-5100-4997-b375-2e60812f4591" containerID="b10d00f07832456c4d17b16ceb1fe7d8e52dfc8871185b619cec4c1be1adacab" exitCode=0 Oct 09 08:29:26 crc kubenswrapper[4872]: I1009 08:29:26.365316 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" event={"ID":"d492de09-5100-4997-b375-2e60812f4591","Type":"ContainerDied","Data":"b10d00f07832456c4d17b16ceb1fe7d8e52dfc8871185b619cec4c1be1adacab"} Oct 09 08:29:27 crc kubenswrapper[4872]: I1009 08:29:27.373907 4872 generic.go:334] "Generic (PLEG): container finished" podID="d492de09-5100-4997-b375-2e60812f4591" containerID="6e0b3146cfaa8c0c9fa3e29cae41615089aa9967e01a979cf4404d8e9ad4d1e3" exitCode=0 Oct 09 08:29:27 crc kubenswrapper[4872]: I1009 08:29:27.373981 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" event={"ID":"d492de09-5100-4997-b375-2e60812f4591","Type":"ContainerDied","Data":"6e0b3146cfaa8c0c9fa3e29cae41615089aa9967e01a979cf4404d8e9ad4d1e3"} Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.674931 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.793688 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util\") pod \"d492de09-5100-4997-b375-2e60812f4591\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.793791 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle\") pod \"d492de09-5100-4997-b375-2e60812f4591\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.793832 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5tbk5\" (UniqueName: \"kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5\") pod \"d492de09-5100-4997-b375-2e60812f4591\" (UID: \"d492de09-5100-4997-b375-2e60812f4591\") " Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.794685 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle" (OuterVolumeSpecName: "bundle") pod "d492de09-5100-4997-b375-2e60812f4591" (UID: "d492de09-5100-4997-b375-2e60812f4591"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.799964 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5" (OuterVolumeSpecName: "kube-api-access-5tbk5") pod "d492de09-5100-4997-b375-2e60812f4591" (UID: "d492de09-5100-4997-b375-2e60812f4591"). InnerVolumeSpecName "kube-api-access-5tbk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.809237 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util" (OuterVolumeSpecName: "util") pod "d492de09-5100-4997-b375-2e60812f4591" (UID: "d492de09-5100-4997-b375-2e60812f4591"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.895057 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.895087 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5tbk5\" (UniqueName: \"kubernetes.io/projected/d492de09-5100-4997-b375-2e60812f4591-kube-api-access-5tbk5\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:28 crc kubenswrapper[4872]: I1009 08:29:28.895098 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d492de09-5100-4997-b375-2e60812f4591-util\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:29 crc kubenswrapper[4872]: I1009 08:29:29.386488 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" event={"ID":"d492de09-5100-4997-b375-2e60812f4591","Type":"ContainerDied","Data":"ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e"} Oct 09 08:29:29 crc kubenswrapper[4872]: I1009 08:29:29.386539 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca44fb891b5a5614f6af2d148d234abba897f0fc377a8fb4101be33c60c8715e" Oct 09 08:29:29 crc kubenswrapper[4872]: I1009 08:29:29.386584 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.615901 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr"] Oct 09 08:29:33 crc kubenswrapper[4872]: E1009 08:29:33.616212 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="util" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.616232 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="util" Oct 09 08:29:33 crc kubenswrapper[4872]: E1009 08:29:33.616251 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="pull" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.616263 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="pull" Oct 09 08:29:33 crc kubenswrapper[4872]: E1009 08:29:33.616288 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="extract" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.616299 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="extract" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.616489 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d492de09-5100-4997-b375-2e60812f4591" containerName="extract" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.617086 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.618547 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-mtwsd" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.620625 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr"] Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.621287 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.632906 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.756195 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b755s\" (UniqueName: \"kubernetes.io/projected/2d597183-b39b-40c1-ac08-ad5467522fc6-kube-api-access-b755s\") pod \"nmstate-operator-858ddd8f98-nwsgr\" (UID: \"2d597183-b39b-40c1-ac08-ad5467522fc6\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.857212 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b755s\" (UniqueName: \"kubernetes.io/projected/2d597183-b39b-40c1-ac08-ad5467522fc6-kube-api-access-b755s\") pod \"nmstate-operator-858ddd8f98-nwsgr\" (UID: \"2d597183-b39b-40c1-ac08-ad5467522fc6\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.885511 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b755s\" (UniqueName: \"kubernetes.io/projected/2d597183-b39b-40c1-ac08-ad5467522fc6-kube-api-access-b755s\") pod \"nmstate-operator-858ddd8f98-nwsgr\" (UID: \"2d597183-b39b-40c1-ac08-ad5467522fc6\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" Oct 09 08:29:33 crc kubenswrapper[4872]: I1009 08:29:33.933186 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" Oct 09 08:29:34 crc kubenswrapper[4872]: I1009 08:29:34.107061 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr"] Oct 09 08:29:34 crc kubenswrapper[4872]: I1009 08:29:34.422971 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" event={"ID":"2d597183-b39b-40c1-ac08-ad5467522fc6","Type":"ContainerStarted","Data":"65e9857242098bee5baf406f83a8efaac38ce583e233749413e2ffc017757482"} Oct 09 08:29:36 crc kubenswrapper[4872]: I1009 08:29:36.441003 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" event={"ID":"2d597183-b39b-40c1-ac08-ad5467522fc6","Type":"ContainerStarted","Data":"328ce8b2d12ebda6364d829b62345ba13028991189c350de8b2007899dc3fd6d"} Oct 09 08:29:36 crc kubenswrapper[4872]: I1009 08:29:36.454520 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-nwsgr" podStartSLOduration=1.655648726 podStartE2EDuration="3.454503766s" podCreationTimestamp="2025-10-09 08:29:33 +0000 UTC" firstStartedPulling="2025-10-09 08:29:34.114367805 +0000 UTC m=+732.304896431" lastFinishedPulling="2025-10-09 08:29:35.913222845 +0000 UTC m=+734.103751471" observedRunningTime="2025-10-09 08:29:36.454473865 +0000 UTC m=+734.645002501" watchObservedRunningTime="2025-10-09 08:29:36.454503766 +0000 UTC m=+734.645032392" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.104468 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.105789 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.110195 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-gm9d9" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.113187 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.125367 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.126382 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.130355 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.136550 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-lgwck"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.142923 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.162275 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.162331 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.195276 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228503 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnx2f\" (UniqueName: \"kubernetes.io/projected/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-kube-api-access-mnx2f\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228574 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5h49\" (UniqueName: \"kubernetes.io/projected/40f6895e-a74d-4ef0-bb3b-9024d4b2db5d-kube-api-access-l5h49\") pod \"nmstate-metrics-fdff9cb8d-fn24v\" (UID: \"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228610 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-dbus-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228629 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-ovs-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228674 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-nmstate-lock\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228708 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbbr\" (UniqueName: \"kubernetes.io/projected/591e63c4-2cfa-478c-a509-e6ad87de2cb5-kube-api-access-hkbbr\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.228729 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.239286 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.240046 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.241822 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.242140 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-j28jd" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.242305 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.251135 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.329748 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-nmstate-lock\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.329824 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-nmstate-lock\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.329883 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxm8r\" (UniqueName: \"kubernetes.io/projected/d9654e2a-2300-47ca-8725-682e7d1cee0a-kube-api-access-wxm8r\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.329943 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbbr\" (UniqueName: \"kubernetes.io/projected/591e63c4-2cfa-478c-a509-e6ad87de2cb5-kube-api-access-hkbbr\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.329967 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330761 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnx2f\" (UniqueName: \"kubernetes.io/projected/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-kube-api-access-mnx2f\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330799 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9654e2a-2300-47ca-8725-682e7d1cee0a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330832 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5h49\" (UniqueName: \"kubernetes.io/projected/40f6895e-a74d-4ef0-bb3b-9024d4b2db5d-kube-api-access-l5h49\") pod \"nmstate-metrics-fdff9cb8d-fn24v\" (UID: \"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330852 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-dbus-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330869 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-ovs-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330885 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9654e2a-2300-47ca-8725-682e7d1cee0a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.330935 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-ovs-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.331196 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/591e63c4-2cfa-478c-a509-e6ad87de2cb5-dbus-socket\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.339123 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.355366 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbbr\" (UniqueName: \"kubernetes.io/projected/591e63c4-2cfa-478c-a509-e6ad87de2cb5-kube-api-access-hkbbr\") pod \"nmstate-handler-lgwck\" (UID: \"591e63c4-2cfa-478c-a509-e6ad87de2cb5\") " pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.356075 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnx2f\" (UniqueName: \"kubernetes.io/projected/d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b-kube-api-access-mnx2f\") pod \"nmstate-webhook-6cdbc54649-ppr2b\" (UID: \"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.359275 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5h49\" (UniqueName: \"kubernetes.io/projected/40f6895e-a74d-4ef0-bb3b-9024d4b2db5d-kube-api-access-l5h49\") pod \"nmstate-metrics-fdff9cb8d-fn24v\" (UID: \"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.417437 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-59f84c84c6-hh826"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.418305 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.431975 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxm8r\" (UniqueName: \"kubernetes.io/projected/d9654e2a-2300-47ca-8725-682e7d1cee0a-kube-api-access-wxm8r\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.432071 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9654e2a-2300-47ca-8725-682e7d1cee0a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.432127 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9654e2a-2300-47ca-8725-682e7d1cee0a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.433585 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/d9654e2a-2300-47ca-8725-682e7d1cee0a-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.437178 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/d9654e2a-2300-47ca-8725-682e7d1cee0a-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.471452 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxm8r\" (UniqueName: \"kubernetes.io/projected/d9654e2a-2300-47ca-8725-682e7d1cee0a-kube-api-access-wxm8r\") pod \"nmstate-console-plugin-6b874cbd85-fz8l7\" (UID: \"d9654e2a-2300-47ca-8725-682e7d1cee0a\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.480983 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.492492 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59f84c84c6-hh826"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.501846 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.505950 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534673 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5pzj\" (UniqueName: \"kubernetes.io/projected/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-kube-api-access-f5pzj\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534718 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534736 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-oauth-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534758 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534784 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-service-ca\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534838 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-trusted-ca-bundle\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.534858 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-oauth-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.559426 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637423 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-trusted-ca-bundle\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637472 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-oauth-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637514 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5pzj\" (UniqueName: \"kubernetes.io/projected/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-kube-api-access-f5pzj\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637541 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637563 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-oauth-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.637632 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-service-ca\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.638744 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-trusted-ca-bundle\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.638856 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.639705 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-oauth-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.639923 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-service-ca\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.644491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-oauth-config\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.645748 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-console-serving-cert\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.656991 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5pzj\" (UniqueName: \"kubernetes.io/projected/3064e5dc-56c7-4ce6-9351-815e6b2ac44c-kube-api-access-f5pzj\") pod \"console-59f84c84c6-hh826\" (UID: \"3064e5dc-56c7-4ce6-9351-815e6b2ac44c\") " pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.750273 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.754317 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b"] Oct 09 08:29:38 crc kubenswrapper[4872]: W1009 08:29:38.767790 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4b82a24_7a85_4e6b_9376_dc4dc2d72e2b.slice/crio-cce2fc5fd9f8a8c20166df73f7f350c12a3e47464a2e71da9178f37084480cb2 WatchSource:0}: Error finding container cce2fc5fd9f8a8c20166df73f7f350c12a3e47464a2e71da9178f37084480cb2: Status 404 returned error can't find the container with id cce2fc5fd9f8a8c20166df73f7f350c12a3e47464a2e71da9178f37084480cb2 Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.931756 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v"] Oct 09 08:29:38 crc kubenswrapper[4872]: I1009 08:29:38.946262 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-59f84c84c6-hh826"] Oct 09 08:29:38 crc kubenswrapper[4872]: W1009 08:29:38.956390 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3064e5dc_56c7_4ce6_9351_815e6b2ac44c.slice/crio-84361874d407eea016e5db4d38e9605e447f850526cd180140824711efc06833 WatchSource:0}: Error finding container 84361874d407eea016e5db4d38e9605e447f850526cd180140824711efc06833: Status 404 returned error can't find the container with id 84361874d407eea016e5db4d38e9605e447f850526cd180140824711efc06833 Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.030260 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7"] Oct 09 08:29:39 crc kubenswrapper[4872]: W1009 08:29:39.035195 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd9654e2a_2300_47ca_8725_682e7d1cee0a.slice/crio-f270d0e19dbfc44dc2941e2e06cd9640112d5e42402f69e6eafd86cdef108a28 WatchSource:0}: Error finding container f270d0e19dbfc44dc2941e2e06cd9640112d5e42402f69e6eafd86cdef108a28: Status 404 returned error can't find the container with id f270d0e19dbfc44dc2941e2e06cd9640112d5e42402f69e6eafd86cdef108a28 Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.460612 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" event={"ID":"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d","Type":"ContainerStarted","Data":"0b80587f1e9d48724489f1f4875b03bd1e46413cc5f5ad1d95b7658961a1e26f"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.462387 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59f84c84c6-hh826" event={"ID":"3064e5dc-56c7-4ce6-9351-815e6b2ac44c","Type":"ContainerStarted","Data":"59c4e13b1e381dbf0285953f455f4b5a534835c7d2df941f757b0092bd1dcb3f"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.462501 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-59f84c84c6-hh826" event={"ID":"3064e5dc-56c7-4ce6-9351-815e6b2ac44c","Type":"ContainerStarted","Data":"84361874d407eea016e5db4d38e9605e447f850526cd180140824711efc06833"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.463642 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lgwck" event={"ID":"591e63c4-2cfa-478c-a509-e6ad87de2cb5","Type":"ContainerStarted","Data":"5d5bdb487b53d5a29b5f39df0acd6188ab4b4df5097ca102e2f3427b9e12d388"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.464969 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" event={"ID":"d9654e2a-2300-47ca-8725-682e7d1cee0a","Type":"ContainerStarted","Data":"f270d0e19dbfc44dc2941e2e06cd9640112d5e42402f69e6eafd86cdef108a28"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.466487 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" event={"ID":"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b","Type":"ContainerStarted","Data":"cce2fc5fd9f8a8c20166df73f7f350c12a3e47464a2e71da9178f37084480cb2"} Oct 09 08:29:39 crc kubenswrapper[4872]: I1009 08:29:39.486776 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-59f84c84c6-hh826" podStartSLOduration=1.486750834 podStartE2EDuration="1.486750834s" podCreationTimestamp="2025-10-09 08:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:29:39.483731056 +0000 UTC m=+737.674259702" watchObservedRunningTime="2025-10-09 08:29:39.486750834 +0000 UTC m=+737.677279460" Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.476747 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-lgwck" event={"ID":"591e63c4-2cfa-478c-a509-e6ad87de2cb5","Type":"ContainerStarted","Data":"246279d11f0d1060e424df311e8f62cb213eac860123fd8e9968d34f2863886b"} Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.477291 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.478092 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" event={"ID":"d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b","Type":"ContainerStarted","Data":"1c9930c116c18d03c1aea11a5c81d454cde4c48370bcad7684ed02acc8a26eec"} Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.478729 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.480424 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" event={"ID":"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d","Type":"ContainerStarted","Data":"d11f31d2ed5538bb49884a10a7810a18971104c39b89e7f8d7e643924cd0a362"} Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.494533 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-lgwck" podStartSLOduration=1.214181573 podStartE2EDuration="3.494507039s" podCreationTimestamp="2025-10-09 08:29:38 +0000 UTC" firstStartedPulling="2025-10-09 08:29:38.558574234 +0000 UTC m=+736.749102860" lastFinishedPulling="2025-10-09 08:29:40.83889971 +0000 UTC m=+739.029428326" observedRunningTime="2025-10-09 08:29:41.490743209 +0000 UTC m=+739.681271845" watchObservedRunningTime="2025-10-09 08:29:41.494507039 +0000 UTC m=+739.685035665" Oct 09 08:29:41 crc kubenswrapper[4872]: I1009 08:29:41.514278 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" podStartSLOduration=1.406864696 podStartE2EDuration="3.514260162s" podCreationTimestamp="2025-10-09 08:29:38 +0000 UTC" firstStartedPulling="2025-10-09 08:29:38.771209536 +0000 UTC m=+736.961738162" lastFinishedPulling="2025-10-09 08:29:40.878605002 +0000 UTC m=+739.069133628" observedRunningTime="2025-10-09 08:29:41.509482423 +0000 UTC m=+739.700011069" watchObservedRunningTime="2025-10-09 08:29:41.514260162 +0000 UTC m=+739.704788778" Oct 09 08:29:42 crc kubenswrapper[4872]: I1009 08:29:42.486822 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" event={"ID":"d9654e2a-2300-47ca-8725-682e7d1cee0a","Type":"ContainerStarted","Data":"fdce63b15e6d5daece04098ac36aea65bab5eb9979eef7b98492bd0494e7834f"} Oct 09 08:29:42 crc kubenswrapper[4872]: I1009 08:29:42.507667 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-fz8l7" podStartSLOduration=1.557467046 podStartE2EDuration="4.507628343s" podCreationTimestamp="2025-10-09 08:29:38 +0000 UTC" firstStartedPulling="2025-10-09 08:29:39.038516644 +0000 UTC m=+737.229045270" lastFinishedPulling="2025-10-09 08:29:41.988677941 +0000 UTC m=+740.179206567" observedRunningTime="2025-10-09 08:29:42.505374468 +0000 UTC m=+740.695903094" watchObservedRunningTime="2025-10-09 08:29:42.507628343 +0000 UTC m=+740.698156979" Oct 09 08:29:43 crc kubenswrapper[4872]: I1009 08:29:43.493881 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" event={"ID":"40f6895e-a74d-4ef0-bb3b-9024d4b2db5d","Type":"ContainerStarted","Data":"b5e16778062405ddd7fbee68aa1a3eb3e547523e8ee5778c47d39c557b20cd44"} Oct 09 08:29:43 crc kubenswrapper[4872]: I1009 08:29:43.514032 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-fn24v" podStartSLOduration=1.535402336 podStartE2EDuration="5.514003693s" podCreationTimestamp="2025-10-09 08:29:38 +0000 UTC" firstStartedPulling="2025-10-09 08:29:38.947093561 +0000 UTC m=+737.137622187" lastFinishedPulling="2025-10-09 08:29:42.925694918 +0000 UTC m=+741.116223544" observedRunningTime="2025-10-09 08:29:43.510057919 +0000 UTC m=+741.700586565" watchObservedRunningTime="2025-10-09 08:29:43.514003693 +0000 UTC m=+741.704532339" Oct 09 08:29:48 crc kubenswrapper[4872]: I1009 08:29:48.571100 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-lgwck" Oct 09 08:29:48 crc kubenswrapper[4872]: I1009 08:29:48.750760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:48 crc kubenswrapper[4872]: I1009 08:29:48.750831 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:48 crc kubenswrapper[4872]: I1009 08:29:48.757380 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.512353 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.512567 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" podUID="09a17519-bda3-4246-bff9-adf452b800b0" containerName="controller-manager" containerID="cri-o://ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92" gracePeriod=30 Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.559569 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-59f84c84c6-hh826" Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.608930 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.622961 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.623194 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerName="route-controller-manager" containerID="cri-o://f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f" gracePeriod=30 Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.906996 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:29:49 crc kubenswrapper[4872]: I1009 08:29:49.976165 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097770 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert\") pod \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097825 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert\") pod \"09a17519-bda3-4246-bff9-adf452b800b0\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097853 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca\") pod \"09a17519-bda3-4246-bff9-adf452b800b0\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097903 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config\") pod \"09a17519-bda3-4246-bff9-adf452b800b0\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097928 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9rhp\" (UniqueName: \"kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp\") pod \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097948 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles\") pod \"09a17519-bda3-4246-bff9-adf452b800b0\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097963 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca\") pod \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.097989 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dx2x\" (UniqueName: \"kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x\") pod \"09a17519-bda3-4246-bff9-adf452b800b0\" (UID: \"09a17519-bda3-4246-bff9-adf452b800b0\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.098023 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config\") pod \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\" (UID: \"953119f6-0fbb-41b4-9901-dfe2dfdcbc98\") " Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.098492 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca" (OuterVolumeSpecName: "client-ca") pod "09a17519-bda3-4246-bff9-adf452b800b0" (UID: "09a17519-bda3-4246-bff9-adf452b800b0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.098571 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config" (OuterVolumeSpecName: "config") pod "09a17519-bda3-4246-bff9-adf452b800b0" (UID: "09a17519-bda3-4246-bff9-adf452b800b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.098707 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "09a17519-bda3-4246-bff9-adf452b800b0" (UID: "09a17519-bda3-4246-bff9-adf452b800b0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.099183 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config" (OuterVolumeSpecName: "config") pod "953119f6-0fbb-41b4-9901-dfe2dfdcbc98" (UID: "953119f6-0fbb-41b4-9901-dfe2dfdcbc98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.102716 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca" (OuterVolumeSpecName: "client-ca") pod "953119f6-0fbb-41b4-9901-dfe2dfdcbc98" (UID: "953119f6-0fbb-41b4-9901-dfe2dfdcbc98"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.103006 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "953119f6-0fbb-41b4-9901-dfe2dfdcbc98" (UID: "953119f6-0fbb-41b4-9901-dfe2dfdcbc98"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.103057 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp" (OuterVolumeSpecName: "kube-api-access-q9rhp") pod "953119f6-0fbb-41b4-9901-dfe2dfdcbc98" (UID: "953119f6-0fbb-41b4-9901-dfe2dfdcbc98"). InnerVolumeSpecName "kube-api-access-q9rhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.103163 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x" (OuterVolumeSpecName: "kube-api-access-6dx2x") pod "09a17519-bda3-4246-bff9-adf452b800b0" (UID: "09a17519-bda3-4246-bff9-adf452b800b0"). InnerVolumeSpecName "kube-api-access-6dx2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.105197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09a17519-bda3-4246-bff9-adf452b800b0" (UID: "09a17519-bda3-4246-bff9-adf452b800b0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199361 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199403 4872 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09a17519-bda3-4246-bff9-adf452b800b0-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199415 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199427 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199436 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9rhp\" (UniqueName: \"kubernetes.io/projected/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-kube-api-access-q9rhp\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199447 4872 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/09a17519-bda3-4246-bff9-adf452b800b0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199458 4872 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-client-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199466 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dx2x\" (UniqueName: \"kubernetes.io/projected/09a17519-bda3-4246-bff9-adf452b800b0-kube-api-access-6dx2x\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.199475 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/953119f6-0fbb-41b4-9901-dfe2dfdcbc98-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.561600 4872 generic.go:334] "Generic (PLEG): container finished" podID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerID="f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f" exitCode=0 Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.561665 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.561675 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" event={"ID":"953119f6-0fbb-41b4-9901-dfe2dfdcbc98","Type":"ContainerDied","Data":"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f"} Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.561700 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j" event={"ID":"953119f6-0fbb-41b4-9901-dfe2dfdcbc98","Type":"ContainerDied","Data":"954442ecb5d7023f091a899fcf7f0d40f18c5500ecd7774d610d9104373f10ab"} Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.561714 4872 scope.go:117] "RemoveContainer" containerID="f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.563056 4872 generic.go:334] "Generic (PLEG): container finished" podID="09a17519-bda3-4246-bff9-adf452b800b0" containerID="ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92" exitCode=0 Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.563108 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.563109 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" event={"ID":"09a17519-bda3-4246-bff9-adf452b800b0","Type":"ContainerDied","Data":"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92"} Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.563204 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-hfxc8" event={"ID":"09a17519-bda3-4246-bff9-adf452b800b0","Type":"ContainerDied","Data":"971e528376523222edae19f0a85b84c9297a639bb77730595f5e9ef79ebb8f27"} Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.580298 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.582784 4872 scope.go:117] "RemoveContainer" containerID="f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.583128 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-hfxc8"] Oct 09 08:29:50 crc kubenswrapper[4872]: E1009 08:29:50.583174 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f\": container with ID starting with f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f not found: ID does not exist" containerID="f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.583217 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f"} err="failed to get container status \"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f\": rpc error: code = NotFound desc = could not find container \"f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f\": container with ID starting with f98d5603569fa325e2e34483f87a2f554baefb20860e2d90f5ca24bc91c40e4f not found: ID does not exist" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.583243 4872 scope.go:117] "RemoveContainer" containerID="ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.588252 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.593857 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-pxn8j"] Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.605474 4872 scope.go:117] "RemoveContainer" containerID="ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92" Oct 09 08:29:50 crc kubenswrapper[4872]: E1009 08:29:50.606455 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92\": container with ID starting with ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92 not found: ID does not exist" containerID="ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92" Oct 09 08:29:50 crc kubenswrapper[4872]: I1009 08:29:50.606487 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92"} err="failed to get container status \"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92\": rpc error: code = NotFound desc = could not find container \"ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92\": container with ID starting with ba7ae88234530ab6cbf257844daeb711eb21272021c7cd0006831a5c99780b92 not found: ID does not exist" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.027269 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-69858c9599-hss7h"] Oct 09 08:29:51 crc kubenswrapper[4872]: E1009 08:29:51.027591 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerName="route-controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.027606 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerName="route-controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: E1009 08:29:51.027632 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09a17519-bda3-4246-bff9-adf452b800b0" containerName="controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.027652 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="09a17519-bda3-4246-bff9-adf452b800b0" containerName="controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.027832 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="09a17519-bda3-4246-bff9-adf452b800b0" containerName="controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.027845 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" containerName="route-controller-manager" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.028386 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.030292 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.030620 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.031364 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.033764 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.036503 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf"] Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.036688 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.037803 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.040140 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.043088 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf"] Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.043603 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.043810 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.043877 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.043922 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.044120 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.044234 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.045956 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.048598 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69858c9599-hss7h"] Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.213967 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-client-ca\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214044 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6x2n\" (UniqueName: \"kubernetes.io/projected/27208101-155a-4bf6-ae7d-c1e50d1cab0d-kube-api-access-q6x2n\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-config\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214095 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z882s\" (UniqueName: \"kubernetes.io/projected/c308548d-c883-4880-8608-7632ab5f0818-kube-api-access-z882s\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214117 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-proxy-ca-bundles\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214133 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-client-ca\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214162 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c308548d-c883-4880-8608-7632ab5f0818-serving-cert\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-config\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.214194 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27208101-155a-4bf6-ae7d-c1e50d1cab0d-serving-cert\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.315689 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-client-ca\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.315798 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-config\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.315833 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c308548d-c883-4880-8608-7632ab5f0818-serving-cert\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.315865 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27208101-155a-4bf6-ae7d-c1e50d1cab0d-serving-cert\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.315915 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-client-ca\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.316002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6x2n\" (UniqueName: \"kubernetes.io/projected/27208101-155a-4bf6-ae7d-c1e50d1cab0d-kube-api-access-q6x2n\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.316032 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-config\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.316089 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z882s\" (UniqueName: \"kubernetes.io/projected/c308548d-c883-4880-8608-7632ab5f0818-kube-api-access-z882s\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.316125 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-proxy-ca-bundles\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.316805 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-client-ca\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.318196 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-client-ca\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.318739 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c308548d-c883-4880-8608-7632ab5f0818-config\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.318971 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-proxy-ca-bundles\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.319148 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/27208101-155a-4bf6-ae7d-c1e50d1cab0d-config\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.322397 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c308548d-c883-4880-8608-7632ab5f0818-serving-cert\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.326227 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/27208101-155a-4bf6-ae7d-c1e50d1cab0d-serving-cert\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.335148 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z882s\" (UniqueName: \"kubernetes.io/projected/c308548d-c883-4880-8608-7632ab5f0818-kube-api-access-z882s\") pod \"route-controller-manager-6d5dc4fb7-st4zf\" (UID: \"c308548d-c883-4880-8608-7632ab5f0818\") " pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.337314 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6x2n\" (UniqueName: \"kubernetes.io/projected/27208101-155a-4bf6-ae7d-c1e50d1cab0d-kube-api-access-q6x2n\") pod \"controller-manager-69858c9599-hss7h\" (UID: \"27208101-155a-4bf6-ae7d-c1e50d1cab0d\") " pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.344600 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.355808 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.556278 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-69858c9599-hss7h"] Oct 09 08:29:51 crc kubenswrapper[4872]: W1009 08:29:51.559124 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27208101_155a_4bf6_ae7d_c1e50d1cab0d.slice/crio-c0b25c52a4ad984e100ab2c1ba3e016801769e159dc64c113c355e3f5eba693f WatchSource:0}: Error finding container c0b25c52a4ad984e100ab2c1ba3e016801769e159dc64c113c355e3f5eba693f: Status 404 returned error can't find the container with id c0b25c52a4ad984e100ab2c1ba3e016801769e159dc64c113c355e3f5eba693f Oct 09 08:29:51 crc kubenswrapper[4872]: I1009 08:29:51.798435 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf"] Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.467742 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09a17519-bda3-4246-bff9-adf452b800b0" path="/var/lib/kubelet/pods/09a17519-bda3-4246-bff9-adf452b800b0/volumes" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.468413 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="953119f6-0fbb-41b4-9901-dfe2dfdcbc98" path="/var/lib/kubelet/pods/953119f6-0fbb-41b4-9901-dfe2dfdcbc98/volumes" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.579190 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" event={"ID":"27208101-155a-4bf6-ae7d-c1e50d1cab0d","Type":"ContainerStarted","Data":"cab9770986ab38ad6e2d2952e57e209b6d657c099abea3950495a9bce483c678"} Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.579242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" event={"ID":"27208101-155a-4bf6-ae7d-c1e50d1cab0d","Type":"ContainerStarted","Data":"c0b25c52a4ad984e100ab2c1ba3e016801769e159dc64c113c355e3f5eba693f"} Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.579610 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.580932 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" event={"ID":"c308548d-c883-4880-8608-7632ab5f0818","Type":"ContainerStarted","Data":"6ce2fb689306e04d4a40613a959a5dead234138f46651563b86f20016d15f16e"} Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.580960 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" event={"ID":"c308548d-c883-4880-8608-7632ab5f0818","Type":"ContainerStarted","Data":"69b2c00234c8c1dc6b8cc02e814ee9c6a08a7f480112164c98c51e7a6935a2c0"} Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.581217 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.584068 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.587457 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.599747 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-69858c9599-hss7h" podStartSLOduration=3.599684441 podStartE2EDuration="3.599684441s" podCreationTimestamp="2025-10-09 08:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:29:52.595436868 +0000 UTC m=+750.785965514" watchObservedRunningTime="2025-10-09 08:29:52.599684441 +0000 UTC m=+750.790213087" Oct 09 08:29:52 crc kubenswrapper[4872]: I1009 08:29:52.636397 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6d5dc4fb7-st4zf" podStartSLOduration=3.636373556 podStartE2EDuration="3.636373556s" podCreationTimestamp="2025-10-09 08:29:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:29:52.632786462 +0000 UTC m=+750.823315088" watchObservedRunningTime="2025-10-09 08:29:52.636373556 +0000 UTC m=+750.826902182" Oct 09 08:29:56 crc kubenswrapper[4872]: I1009 08:29:56.676954 4872 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 09 08:29:58 crc kubenswrapper[4872]: I1009 08:29:58.508579 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-ppr2b" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.127264 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h"] Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.128622 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.134446 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.134716 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.143872 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h"] Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.231871 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.231930 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpjmt\" (UniqueName: \"kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.231965 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.334565 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.334630 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpjmt\" (UniqueName: \"kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.334684 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.335542 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.340668 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.356879 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpjmt\" (UniqueName: \"kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt\") pod \"collect-profiles-29333310-cjj7h\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.486314 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:00 crc kubenswrapper[4872]: I1009 08:30:00.872598 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h"] Oct 09 08:30:00 crc kubenswrapper[4872]: W1009 08:30:00.879425 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1eb69287_4970_49c0_a89b_28372b40743c.slice/crio-d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7 WatchSource:0}: Error finding container d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7: Status 404 returned error can't find the container with id d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7 Oct 09 08:30:01 crc kubenswrapper[4872]: I1009 08:30:01.634907 4872 generic.go:334] "Generic (PLEG): container finished" podID="1eb69287-4970-49c0-a89b-28372b40743c" containerID="c9cb5738f683c120377f608e0df0a20b4d9b24150d71ec4cfe22a12a724b0fc5" exitCode=0 Oct 09 08:30:01 crc kubenswrapper[4872]: I1009 08:30:01.635000 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" event={"ID":"1eb69287-4970-49c0-a89b-28372b40743c","Type":"ContainerDied","Data":"c9cb5738f683c120377f608e0df0a20b4d9b24150d71ec4cfe22a12a724b0fc5"} Oct 09 08:30:01 crc kubenswrapper[4872]: I1009 08:30:01.635239 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" event={"ID":"1eb69287-4970-49c0-a89b-28372b40743c","Type":"ContainerStarted","Data":"d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7"} Oct 09 08:30:02 crc kubenswrapper[4872]: I1009 08:30:02.943776 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.069541 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lpjmt\" (UniqueName: \"kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt\") pod \"1eb69287-4970-49c0-a89b-28372b40743c\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.069593 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume\") pod \"1eb69287-4970-49c0-a89b-28372b40743c\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.069699 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume\") pod \"1eb69287-4970-49c0-a89b-28372b40743c\" (UID: \"1eb69287-4970-49c0-a89b-28372b40743c\") " Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.070810 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume" (OuterVolumeSpecName: "config-volume") pod "1eb69287-4970-49c0-a89b-28372b40743c" (UID: "1eb69287-4970-49c0-a89b-28372b40743c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.075202 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt" (OuterVolumeSpecName: "kube-api-access-lpjmt") pod "1eb69287-4970-49c0-a89b-28372b40743c" (UID: "1eb69287-4970-49c0-a89b-28372b40743c"). InnerVolumeSpecName "kube-api-access-lpjmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.075757 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1eb69287-4970-49c0-a89b-28372b40743c" (UID: "1eb69287-4970-49c0-a89b-28372b40743c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.171394 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lpjmt\" (UniqueName: \"kubernetes.io/projected/1eb69287-4970-49c0-a89b-28372b40743c-kube-api-access-lpjmt\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.171451 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1eb69287-4970-49c0-a89b-28372b40743c-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.171465 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1eb69287-4970-49c0-a89b-28372b40743c-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.646410 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" event={"ID":"1eb69287-4970-49c0-a89b-28372b40743c","Type":"ContainerDied","Data":"d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7"} Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.646840 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d860451be3daa064560f724adbd5071ebc74db58bcf554cb6a23ced5d3f2a7c7" Oct 09 08:30:03 crc kubenswrapper[4872]: I1009 08:30:03.646490 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h" Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.157569 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.158108 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.158160 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.158818 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.158886 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd" gracePeriod=600 Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.687921 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd" exitCode=0 Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.688101 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd"} Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.688646 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41"} Oct 09 08:30:08 crc kubenswrapper[4872]: I1009 08:30:08.688670 4872 scope.go:117] "RemoveContainer" containerID="3c464a6cd1741b02f33a6812243606f46a3bba188820d0e683c67770e869404c" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.239341 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm"] Oct 09 08:30:12 crc kubenswrapper[4872]: E1009 08:30:12.241675 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1eb69287-4970-49c0-a89b-28372b40743c" containerName="collect-profiles" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.241888 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1eb69287-4970-49c0-a89b-28372b40743c" containerName="collect-profiles" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.242201 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1eb69287-4970-49c0-a89b-28372b40743c" containerName="collect-profiles" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.243577 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.246618 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.262176 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm"] Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.401809 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzx4k\" (UniqueName: \"kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.401870 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.401928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.502606 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.502737 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzx4k\" (UniqueName: \"kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.502773 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.503347 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.503349 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.522378 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzx4k\" (UniqueName: \"kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:12 crc kubenswrapper[4872]: I1009 08:30:12.609402 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:13 crc kubenswrapper[4872]: I1009 08:30:13.051777 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm"] Oct 09 08:30:13 crc kubenswrapper[4872]: I1009 08:30:13.723009 4872 generic.go:334] "Generic (PLEG): container finished" podID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerID="a4456c506007d44b42f64c3a805424ecf737a5e11d99aa7e3c65749cc160a8a2" exitCode=0 Oct 09 08:30:13 crc kubenswrapper[4872]: I1009 08:30:13.723178 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" event={"ID":"fbdc30ed-ca91-4dc0-b908-a6781a0857e4","Type":"ContainerDied","Data":"a4456c506007d44b42f64c3a805424ecf737a5e11d99aa7e3c65749cc160a8a2"} Oct 09 08:30:13 crc kubenswrapper[4872]: I1009 08:30:13.723777 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" event={"ID":"fbdc30ed-ca91-4dc0-b908-a6781a0857e4","Type":"ContainerStarted","Data":"29ae1e9dc4a351d581a0e1e07abece7635275ecbf81da37db0e981aa0306da42"} Oct 09 08:30:14 crc kubenswrapper[4872]: I1009 08:30:14.668411 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-4gkkd" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerName="console" containerID="cri-o://4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562" gracePeriod=15 Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.250410 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-4gkkd_2b9767b1-3531-4558-a6d7-55cae92ad568/console/0.log" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.250876 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.341911 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.341986 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.343630 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.343729 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.343755 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.343807 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.343833 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snct7\" (UniqueName: \"kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7\") pod \"2b9767b1-3531-4558-a6d7-55cae92ad568\" (UID: \"2b9767b1-3531-4558-a6d7-55cae92ad568\") " Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.344605 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.344627 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca" (OuterVolumeSpecName: "service-ca") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.344664 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config" (OuterVolumeSpecName: "console-config") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.345223 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.349931 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.350557 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7" (OuterVolumeSpecName: "kube-api-access-snct7") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "kube-api-access-snct7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.355624 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "2b9767b1-3531-4558-a6d7-55cae92ad568" (UID: "2b9767b1-3531-4558-a6d7-55cae92ad568"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445400 4872 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445439 4872 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445466 4872 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-service-ca\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445475 4872 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445484 4872 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/2b9767b1-3531-4558-a6d7-55cae92ad568-console-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445492 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snct7\" (UniqueName: \"kubernetes.io/projected/2b9767b1-3531-4558-a6d7-55cae92ad568-kube-api-access-snct7\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.445505 4872 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/2b9767b1-3531-4558-a6d7-55cae92ad568-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.596504 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:15 crc kubenswrapper[4872]: E1009 08:30:15.597040 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerName="console" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.597078 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerName="console" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.597340 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerName="console" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.599163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.605572 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.739478 4872 generic.go:334] "Generic (PLEG): container finished" podID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerID="323c3b469feaa9d2d38bc3aad655232cd4e2246df079bff0957ae7d31c32e7f6" exitCode=0 Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.739514 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" event={"ID":"fbdc30ed-ca91-4dc0-b908-a6781a0857e4","Type":"ContainerDied","Data":"323c3b469feaa9d2d38bc3aad655232cd4e2246df079bff0957ae7d31c32e7f6"} Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741098 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-4gkkd_2b9767b1-3531-4558-a6d7-55cae92ad568/console/0.log" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741129 4872 generic.go:334] "Generic (PLEG): container finished" podID="2b9767b1-3531-4558-a6d7-55cae92ad568" containerID="4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562" exitCode=2 Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4gkkd" event={"ID":"2b9767b1-3531-4558-a6d7-55cae92ad568","Type":"ContainerDied","Data":"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562"} Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741164 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-4gkkd" event={"ID":"2b9767b1-3531-4558-a6d7-55cae92ad568","Type":"ContainerDied","Data":"83d134f548d7652d63c0950088be313e24bc3c1d7c00e7efeeee79728b673d0b"} Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741183 4872 scope.go:117] "RemoveContainer" containerID="4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.741232 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-4gkkd" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.749114 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.749203 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmhc8\" (UniqueName: \"kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.749233 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.768204 4872 scope.go:117] "RemoveContainer" containerID="4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562" Oct 09 08:30:15 crc kubenswrapper[4872]: E1009 08:30:15.768626 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562\": container with ID starting with 4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562 not found: ID does not exist" containerID="4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.768681 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562"} err="failed to get container status \"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562\": rpc error: code = NotFound desc = could not find container \"4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562\": container with ID starting with 4d023981480e18d11e2120c0365cd20ec7ce96eaa9d20784559144953e05e562 not found: ID does not exist" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.785720 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.793670 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-4gkkd"] Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.850810 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.850854 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmhc8\" (UniqueName: \"kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.850891 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.851387 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.851396 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.875410 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmhc8\" (UniqueName: \"kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8\") pod \"redhat-operators-nxsts\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:15 crc kubenswrapper[4872]: I1009 08:30:15.930336 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.367835 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:16 crc kubenswrapper[4872]: W1009 08:30:16.374692 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4206758_e4d1_4ac8_bf5e_1e6b28de0942.slice/crio-6b6975e63fe61fcaa1f9cceee7e49aad72dc83ce3efeb3e8a844afffa184dba0 WatchSource:0}: Error finding container 6b6975e63fe61fcaa1f9cceee7e49aad72dc83ce3efeb3e8a844afffa184dba0: Status 404 returned error can't find the container with id 6b6975e63fe61fcaa1f9cceee7e49aad72dc83ce3efeb3e8a844afffa184dba0 Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.469414 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b9767b1-3531-4558-a6d7-55cae92ad568" path="/var/lib/kubelet/pods/2b9767b1-3531-4558-a6d7-55cae92ad568/volumes" Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.751301 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerID="c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6" exitCode=0 Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.751374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerDied","Data":"c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6"} Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.751399 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerStarted","Data":"6b6975e63fe61fcaa1f9cceee7e49aad72dc83ce3efeb3e8a844afffa184dba0"} Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.754694 4872 generic.go:334] "Generic (PLEG): container finished" podID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerID="ac9240afd9ad1085b048a59665172004d557e97343242a1dfe20ed5e66101fee" exitCode=0 Oct 09 08:30:16 crc kubenswrapper[4872]: I1009 08:30:16.754723 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" event={"ID":"fbdc30ed-ca91-4dc0-b908-a6781a0857e4","Type":"ContainerDied","Data":"ac9240afd9ad1085b048a59665172004d557e97343242a1dfe20ed5e66101fee"} Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.105477 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.285256 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle\") pod \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.285351 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzx4k\" (UniqueName: \"kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k\") pod \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.285415 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util\") pod \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\" (UID: \"fbdc30ed-ca91-4dc0-b908-a6781a0857e4\") " Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.286313 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle" (OuterVolumeSpecName: "bundle") pod "fbdc30ed-ca91-4dc0-b908-a6781a0857e4" (UID: "fbdc30ed-ca91-4dc0-b908-a6781a0857e4"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.287145 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.292981 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k" (OuterVolumeSpecName: "kube-api-access-hzx4k") pod "fbdc30ed-ca91-4dc0-b908-a6781a0857e4" (UID: "fbdc30ed-ca91-4dc0-b908-a6781a0857e4"). InnerVolumeSpecName "kube-api-access-hzx4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.303570 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util" (OuterVolumeSpecName: "util") pod "fbdc30ed-ca91-4dc0-b908-a6781a0857e4" (UID: "fbdc30ed-ca91-4dc0-b908-a6781a0857e4"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.388779 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzx4k\" (UniqueName: \"kubernetes.io/projected/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-kube-api-access-hzx4k\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.388814 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fbdc30ed-ca91-4dc0-b908-a6781a0857e4-util\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.769753 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" event={"ID":"fbdc30ed-ca91-4dc0-b908-a6781a0857e4","Type":"ContainerDied","Data":"29ae1e9dc4a351d581a0e1e07abece7635275ecbf81da37db0e981aa0306da42"} Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.769777 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.769796 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29ae1e9dc4a351d581a0e1e07abece7635275ecbf81da37db0e981aa0306da42" Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.772000 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerID="33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f" exitCode=0 Oct 09 08:30:18 crc kubenswrapper[4872]: I1009 08:30:18.772073 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerDied","Data":"33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f"} Oct 09 08:30:19 crc kubenswrapper[4872]: I1009 08:30:19.779145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerStarted","Data":"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5"} Oct 09 08:30:19 crc kubenswrapper[4872]: I1009 08:30:19.801727 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nxsts" podStartSLOduration=2.364336716 podStartE2EDuration="4.801695849s" podCreationTimestamp="2025-10-09 08:30:15 +0000 UTC" firstStartedPulling="2025-10-09 08:30:16.753715432 +0000 UTC m=+774.944244058" lastFinishedPulling="2025-10-09 08:30:19.191074565 +0000 UTC m=+777.381603191" observedRunningTime="2025-10-09 08:30:19.797467506 +0000 UTC m=+777.987996192" watchObservedRunningTime="2025-10-09 08:30:19.801695849 +0000 UTC m=+777.992224475" Oct 09 08:30:25 crc kubenswrapper[4872]: I1009 08:30:25.930745 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:25 crc kubenswrapper[4872]: I1009 08:30:25.931116 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:25 crc kubenswrapper[4872]: I1009 08:30:25.969538 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:26 crc kubenswrapper[4872]: I1009 08:30:26.874131 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.103595 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs"] Oct 09 08:30:28 crc kubenswrapper[4872]: E1009 08:30:28.103855 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="util" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.103867 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="util" Oct 09 08:30:28 crc kubenswrapper[4872]: E1009 08:30:28.103877 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="extract" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.103883 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="extract" Oct 09 08:30:28 crc kubenswrapper[4872]: E1009 08:30:28.103899 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="pull" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.103905 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="pull" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.103994 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fbdc30ed-ca91-4dc0-b908-a6781a0857e4" containerName="extract" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.104335 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107039 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107186 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107301 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tbbvt" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107354 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxksg\" (UniqueName: \"kubernetes.io/projected/75ddc562-e0fb-4c1c-b98d-eb754c97b989-kube-api-access-gxksg\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107395 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-webhook-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107430 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-apiservice-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107457 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.107820 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.162334 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs"] Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.208520 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-apiservice-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.208597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxksg\" (UniqueName: \"kubernetes.io/projected/75ddc562-e0fb-4c1c-b98d-eb754c97b989-kube-api-access-gxksg\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.208624 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-webhook-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.214575 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-apiservice-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.218584 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/75ddc562-e0fb-4c1c-b98d-eb754c97b989-webhook-cert\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.224668 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxksg\" (UniqueName: \"kubernetes.io/projected/75ddc562-e0fb-4c1c-b98d-eb754c97b989-kube-api-access-gxksg\") pod \"metallb-operator-controller-manager-6d774594fc-m77qs\" (UID: \"75ddc562-e0fb-4c1c-b98d-eb754c97b989\") " pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.359938 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq"] Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.360784 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.362993 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.363099 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.364468 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-x9q9j" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.398174 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq"] Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.418066 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.512222 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pmcq\" (UniqueName: \"kubernetes.io/projected/c49185d6-a699-454b-8e58-88d86a55b694-kube-api-access-7pmcq\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.512262 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-apiservice-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.512283 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-webhook-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.618601 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pmcq\" (UniqueName: \"kubernetes.io/projected/c49185d6-a699-454b-8e58-88d86a55b694-kube-api-access-7pmcq\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.618675 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-apiservice-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.618708 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-webhook-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.628221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-webhook-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.628803 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c49185d6-a699-454b-8e58-88d86a55b694-apiservice-cert\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.639070 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pmcq\" (UniqueName: \"kubernetes.io/projected/c49185d6-a699-454b-8e58-88d86a55b694-kube-api-access-7pmcq\") pod \"metallb-operator-webhook-server-6f4fb54f85-5jlsq\" (UID: \"c49185d6-a699-454b-8e58-88d86a55b694\") " pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.677633 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:28 crc kubenswrapper[4872]: I1009 08:30:28.893408 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs"] Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.083184 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq"] Oct 09 08:30:29 crc kubenswrapper[4872]: W1009 08:30:29.094611 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc49185d6_a699_454b_8e58_88d86a55b694.slice/crio-ec9f02389432a58cf7fa22794b43bca2889877c9d739dced392b6c27f589b236 WatchSource:0}: Error finding container ec9f02389432a58cf7fa22794b43bca2889877c9d739dced392b6c27f589b236: Status 404 returned error can't find the container with id ec9f02389432a58cf7fa22794b43bca2889877c9d739dced392b6c27f589b236 Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.174143 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.175454 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nxsts" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="registry-server" containerID="cri-o://353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5" gracePeriod=2 Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.547539 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.634141 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities\") pod \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.634188 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmhc8\" (UniqueName: \"kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8\") pod \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.634270 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content\") pod \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\" (UID: \"a4206758-e4d1-4ac8-bf5e-1e6b28de0942\") " Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.635353 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities" (OuterVolumeSpecName: "utilities") pod "a4206758-e4d1-4ac8-bf5e-1e6b28de0942" (UID: "a4206758-e4d1-4ac8-bf5e-1e6b28de0942"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.638376 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8" (OuterVolumeSpecName: "kube-api-access-pmhc8") pod "a4206758-e4d1-4ac8-bf5e-1e6b28de0942" (UID: "a4206758-e4d1-4ac8-bf5e-1e6b28de0942"). InnerVolumeSpecName "kube-api-access-pmhc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.713016 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a4206758-e4d1-4ac8-bf5e-1e6b28de0942" (UID: "a4206758-e4d1-4ac8-bf5e-1e6b28de0942"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.735920 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.735951 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.735961 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmhc8\" (UniqueName: \"kubernetes.io/projected/a4206758-e4d1-4ac8-bf5e-1e6b28de0942-kube-api-access-pmhc8\") on node \"crc\" DevicePath \"\"" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.846659 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" event={"ID":"75ddc562-e0fb-4c1c-b98d-eb754c97b989","Type":"ContainerStarted","Data":"818a78d2ad724b06d31197f529d0c0e4cb09a9d730a5d28118b0c42f61823c9d"} Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.849849 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerID="353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5" exitCode=0 Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.849907 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerDied","Data":"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5"} Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.849931 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nxsts" event={"ID":"a4206758-e4d1-4ac8-bf5e-1e6b28de0942","Type":"ContainerDied","Data":"6b6975e63fe61fcaa1f9cceee7e49aad72dc83ce3efeb3e8a844afffa184dba0"} Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.849952 4872 scope.go:117] "RemoveContainer" containerID="353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.850096 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nxsts" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.854009 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" event={"ID":"c49185d6-a699-454b-8e58-88d86a55b694","Type":"ContainerStarted","Data":"ec9f02389432a58cf7fa22794b43bca2889877c9d739dced392b6c27f589b236"} Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.876734 4872 scope.go:117] "RemoveContainer" containerID="33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.882810 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.895355 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nxsts"] Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.896495 4872 scope.go:117] "RemoveContainer" containerID="c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.909082 4872 scope.go:117] "RemoveContainer" containerID="353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5" Oct 09 08:30:29 crc kubenswrapper[4872]: E1009 08:30:29.909539 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5\": container with ID starting with 353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5 not found: ID does not exist" containerID="353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.909572 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5"} err="failed to get container status \"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5\": rpc error: code = NotFound desc = could not find container \"353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5\": container with ID starting with 353c6c567205aa345cdf2bdae8f182139d2983a6746b8fad9539d38164cad9a5 not found: ID does not exist" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.909595 4872 scope.go:117] "RemoveContainer" containerID="33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f" Oct 09 08:30:29 crc kubenswrapper[4872]: E1009 08:30:29.909908 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f\": container with ID starting with 33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f not found: ID does not exist" containerID="33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.909928 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f"} err="failed to get container status \"33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f\": rpc error: code = NotFound desc = could not find container \"33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f\": container with ID starting with 33a46b46fc9d6518750e341634ad376dad39f73f4a5665ba0f679b6db8d6500f not found: ID does not exist" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.909940 4872 scope.go:117] "RemoveContainer" containerID="c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6" Oct 09 08:30:29 crc kubenswrapper[4872]: E1009 08:30:29.910212 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6\": container with ID starting with c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6 not found: ID does not exist" containerID="c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6" Oct 09 08:30:29 crc kubenswrapper[4872]: I1009 08:30:29.910232 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6"} err="failed to get container status \"c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6\": rpc error: code = NotFound desc = could not find container \"c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6\": container with ID starting with c73148c9ba1be827da76ba6aeb7b99dc9ece2528baa4cd2a90f07eedfce2b9b6 not found: ID does not exist" Oct 09 08:30:30 crc kubenswrapper[4872]: I1009 08:30:30.474972 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" path="/var/lib/kubelet/pods/a4206758-e4d1-4ac8-bf5e-1e6b28de0942/volumes" Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.882651 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" event={"ID":"c49185d6-a699-454b-8e58-88d86a55b694","Type":"ContainerStarted","Data":"7620f200c72150971765038a4f385c4b37220700a9aa4b73b4f1d1a6defd160e"} Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.883304 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.884890 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" event={"ID":"75ddc562-e0fb-4c1c-b98d-eb754c97b989","Type":"ContainerStarted","Data":"2e0846f6d49e8a57d379ac46db845dc60a5933ce8201af4d16e3e693cc4f2469"} Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.885054 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.960175 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" podStartSLOduration=1.6175404530000002 podStartE2EDuration="5.960152046s" podCreationTimestamp="2025-10-09 08:30:28 +0000 UTC" firstStartedPulling="2025-10-09 08:30:28.901721623 +0000 UTC m=+787.092250249" lastFinishedPulling="2025-10-09 08:30:33.244333206 +0000 UTC m=+791.434861842" observedRunningTime="2025-10-09 08:30:33.957087167 +0000 UTC m=+792.147615803" watchObservedRunningTime="2025-10-09 08:30:33.960152046 +0000 UTC m=+792.150680702" Oct 09 08:30:33 crc kubenswrapper[4872]: I1009 08:30:33.962376 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" podStartSLOduration=1.7910177950000001 podStartE2EDuration="5.962361189s" podCreationTimestamp="2025-10-09 08:30:28 +0000 UTC" firstStartedPulling="2025-10-09 08:30:29.098363486 +0000 UTC m=+787.288892112" lastFinishedPulling="2025-10-09 08:30:33.26970687 +0000 UTC m=+791.460235506" observedRunningTime="2025-10-09 08:30:33.923220268 +0000 UTC m=+792.113748924" watchObservedRunningTime="2025-10-09 08:30:33.962361189 +0000 UTC m=+792.152889855" Oct 09 08:30:48 crc kubenswrapper[4872]: I1009 08:30:48.684543 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f4fb54f85-5jlsq" Oct 09 08:31:08 crc kubenswrapper[4872]: I1009 08:31:08.422564 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d774594fc-m77qs" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.217710 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg"] Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.217963 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="extract-content" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.217982 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="extract-content" Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.218012 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="extract-utilities" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.218022 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="extract-utilities" Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.218032 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="registry-server" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.218040 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="registry-server" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.218148 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4206758-e4d1-4ac8-bf5e-1e6b28de0942" containerName="registry-server" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.218489 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.223480 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.223584 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bstr8" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.239375 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg"] Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.258433 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hd8n4"] Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.260583 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: W1009 08:31:09.266794 4872 reflector.go:561] object-"metallb-system"/"frr-k8s-certs-secret": failed to list *v1.Secret: secrets "frr-k8s-certs-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.266846 4872 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-k8s-certs-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"frr-k8s-certs-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:31:09 crc kubenswrapper[4872]: W1009 08:31:09.266847 4872 reflector.go:561] object-"metallb-system"/"frr-startup": failed to list *v1.ConfigMap: configmaps "frr-startup" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "metallb-system": no relationship found between node 'crc' and this object Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.266900 4872 reflector.go:158] "Unhandled Error" err="object-\"metallb-system\"/\"frr-startup\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"frr-startup\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"metallb-system\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.359737 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-reloader\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360294 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-conf\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360318 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360343 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hszzv\" (UniqueName: \"kubernetes.io/projected/381ccfd5-05cf-4592-933e-128d714bf0a6-kube-api-access-hszzv\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360379 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics-certs\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360573 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-sockets\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360617 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-startup\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.360829 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldwlj\" (UniqueName: \"kubernetes.io/projected/94c5800f-97ae-4482-a959-d459cbdfdbaf-kube-api-access-ldwlj\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.374947 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-b52wr"] Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.375861 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.380833 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.380854 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-m8w7v" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.381070 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.381165 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.408336 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-2hm62"] Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.409163 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.410994 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.437304 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2hm62"] Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461750 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-reloader\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461806 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-conf\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461831 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461857 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hszzv\" (UniqueName: \"kubernetes.io/projected/381ccfd5-05cf-4592-933e-128d714bf0a6-kube-api-access-hszzv\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461892 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics-certs\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461917 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461937 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-sockets\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.461966 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-startup\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.462015 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldwlj\" (UniqueName: \"kubernetes.io/projected/94c5800f-97ae-4482-a959-d459cbdfdbaf-kube-api-access-ldwlj\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.462278 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-conf\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.462468 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-reloader\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.462715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.462783 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-sockets\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.462821 4872 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.462876 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert podName:381ccfd5-05cf-4592-933e-128d714bf0a6 nodeName:}" failed. No retries permitted until 2025-10-09 08:31:09.962858319 +0000 UTC m=+828.153386965 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert") pod "frr-k8s-webhook-server-64bf5d555-9zqhg" (UID: "381ccfd5-05cf-4592-933e-128d714bf0a6") : secret "frr-k8s-webhook-server-cert" not found Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.488472 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldwlj\" (UniqueName: \"kubernetes.io/projected/94c5800f-97ae-4482-a959-d459cbdfdbaf-kube-api-access-ldwlj\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.496565 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hszzv\" (UniqueName: \"kubernetes.io/projected/381ccfd5-05cf-4592-933e-128d714bf0a6-kube-api-access-hszzv\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.563399 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-metrics-certs\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.564580 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssddb\" (UniqueName: \"kubernetes.io/projected/76fe700a-cada-4802-91e5-1bd3522ab565-kube-api-access-ssddb\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.564621 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-cert\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.564717 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-metrics-certs\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.564738 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.564968 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p8ck\" (UniqueName: \"kubernetes.io/projected/fadbcba0-ee7c-4e6b-8134-b60655294553-kube-api-access-7p8ck\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.565029 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76fe700a-cada-4802-91e5-1bd3522ab565-metallb-excludel2\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.666220 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssddb\" (UniqueName: \"kubernetes.io/projected/76fe700a-cada-4802-91e5-1bd3522ab565-kube-api-access-ssddb\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.666552 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-cert\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.666721 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-metrics-certs\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.666848 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.667049 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p8ck\" (UniqueName: \"kubernetes.io/projected/fadbcba0-ee7c-4e6b-8134-b60655294553-kube-api-access-7p8ck\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.667177 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76fe700a-cada-4802-91e5-1bd3522ab565-metallb-excludel2\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.667318 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-metrics-certs\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.667068 4872 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 09 08:31:09 crc kubenswrapper[4872]: E1009 08:31:09.667712 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist podName:76fe700a-cada-4802-91e5-1bd3522ab565 nodeName:}" failed. No retries permitted until 2025-10-09 08:31:10.16769081 +0000 UTC m=+828.358219436 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist") pod "speaker-b52wr" (UID: "76fe700a-cada-4802-91e5-1bd3522ab565") : secret "metallb-memberlist" not found Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.668109 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/76fe700a-cada-4802-91e5-1bd3522ab565-metallb-excludel2\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.678908 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-cert\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.679201 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fadbcba0-ee7c-4e6b-8134-b60655294553-metrics-certs\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.682094 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssddb\" (UniqueName: \"kubernetes.io/projected/76fe700a-cada-4802-91e5-1bd3522ab565-kube-api-access-ssddb\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.686132 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-metrics-certs\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.692849 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p8ck\" (UniqueName: \"kubernetes.io/projected/fadbcba0-ee7c-4e6b-8134-b60655294553-kube-api-access-7p8ck\") pod \"controller-68d546b9d8-2hm62\" (UID: \"fadbcba0-ee7c-4e6b-8134-b60655294553\") " pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.723252 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.972312 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:09 crc kubenswrapper[4872]: I1009 08:31:09.977595 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/381ccfd5-05cf-4592-933e-128d714bf0a6-cert\") pod \"frr-k8s-webhook-server-64bf5d555-9zqhg\" (UID: \"381ccfd5-05cf-4592-933e-128d714bf0a6\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.084851 4872 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.096392 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/94c5800f-97ae-4482-a959-d459cbdfdbaf-metrics-certs\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.130356 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-2hm62"] Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.138527 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.174796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:10 crc kubenswrapper[4872]: E1009 08:31:10.174972 4872 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 09 08:31:10 crc kubenswrapper[4872]: E1009 08:31:10.175065 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist podName:76fe700a-cada-4802-91e5-1bd3522ab565 nodeName:}" failed. No retries permitted until 2025-10-09 08:31:11.175040453 +0000 UTC m=+829.365569079 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist") pod "speaker-b52wr" (UID: "76fe700a-cada-4802-91e5-1bd3522ab565") : secret "metallb-memberlist" not found Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.216993 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.223851 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/94c5800f-97ae-4482-a959-d459cbdfdbaf-frr-startup\") pod \"frr-k8s-hd8n4\" (UID: \"94c5800f-97ae-4482-a959-d459cbdfdbaf\") " pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.478223 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:10 crc kubenswrapper[4872]: I1009 08:31:10.967904 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg"] Oct 09 08:31:10 crc kubenswrapper[4872]: W1009 08:31:10.973417 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod381ccfd5_05cf_4592_933e_128d714bf0a6.slice/crio-624a9ba21b3a035410d437841d29b28525db9eac404ec8edb698a14b0a0d371e WatchSource:0}: Error finding container 624a9ba21b3a035410d437841d29b28525db9eac404ec8edb698a14b0a0d371e: Status 404 returned error can't find the container with id 624a9ba21b3a035410d437841d29b28525db9eac404ec8edb698a14b0a0d371e Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.105315 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" event={"ID":"381ccfd5-05cf-4592-933e-128d714bf0a6","Type":"ContainerStarted","Data":"624a9ba21b3a035410d437841d29b28525db9eac404ec8edb698a14b0a0d371e"} Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.106669 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"694c277eaaea13297c6783c6a88680957be9b5442880b0d49efa5d9668515ba5"} Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.108365 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2hm62" event={"ID":"fadbcba0-ee7c-4e6b-8134-b60655294553","Type":"ContainerStarted","Data":"abb237f0680bc3d9f0670665e5da3fe3c33dc6f860676cb1c816d2031d535cd3"} Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.108396 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2hm62" event={"ID":"fadbcba0-ee7c-4e6b-8134-b60655294553","Type":"ContainerStarted","Data":"78f80333d5de9f429a5b64809b19fe8fa6641a7af3c9cb37e9d53a085ab975bf"} Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.108406 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-2hm62" event={"ID":"fadbcba0-ee7c-4e6b-8134-b60655294553","Type":"ContainerStarted","Data":"fb177dc8a490ad7532fcdecd2eb3958404e367d1ba889e1556a8f2d739b3912e"} Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.108506 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.130024 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-2hm62" podStartSLOduration=2.130006425 podStartE2EDuration="2.130006425s" podCreationTimestamp="2025-10-09 08:31:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:31:11.125303649 +0000 UTC m=+829.315832285" watchObservedRunningTime="2025-10-09 08:31:11.130006425 +0000 UTC m=+829.320535051" Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.188153 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.199560 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/76fe700a-cada-4802-91e5-1bd3522ab565-memberlist\") pod \"speaker-b52wr\" (UID: \"76fe700a-cada-4802-91e5-1bd3522ab565\") " pod="metallb-system/speaker-b52wr" Oct 09 08:31:11 crc kubenswrapper[4872]: I1009 08:31:11.491273 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-b52wr" Oct 09 08:31:11 crc kubenswrapper[4872]: W1009 08:31:11.515866 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod76fe700a_cada_4802_91e5_1bd3522ab565.slice/crio-ab3ba6ac9964157af0f9a5bc852ea376a11368fd4a8236cd72545e2439c3a92d WatchSource:0}: Error finding container ab3ba6ac9964157af0f9a5bc852ea376a11368fd4a8236cd72545e2439c3a92d: Status 404 returned error can't find the container with id ab3ba6ac9964157af0f9a5bc852ea376a11368fd4a8236cd72545e2439c3a92d Oct 09 08:31:12 crc kubenswrapper[4872]: I1009 08:31:12.119033 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-b52wr" event={"ID":"76fe700a-cada-4802-91e5-1bd3522ab565","Type":"ContainerStarted","Data":"49ba8292b107a4bdd62b1d269ae6ca5781b8913fadded93a3d806a0b2c157e41"} Oct 09 08:31:12 crc kubenswrapper[4872]: I1009 08:31:12.119073 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-b52wr" event={"ID":"76fe700a-cada-4802-91e5-1bd3522ab565","Type":"ContainerStarted","Data":"960d24ba46cf986470c26807499b1a30949d3710f5c2a592300082a46994f2c7"} Oct 09 08:31:12 crc kubenswrapper[4872]: I1009 08:31:12.119082 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-b52wr" event={"ID":"76fe700a-cada-4802-91e5-1bd3522ab565","Type":"ContainerStarted","Data":"ab3ba6ac9964157af0f9a5bc852ea376a11368fd4a8236cd72545e2439c3a92d"} Oct 09 08:31:12 crc kubenswrapper[4872]: I1009 08:31:12.119173 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-b52wr" Oct 09 08:31:12 crc kubenswrapper[4872]: I1009 08:31:12.143379 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-b52wr" podStartSLOduration=3.143363793 podStartE2EDuration="3.143363793s" podCreationTimestamp="2025-10-09 08:31:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:31:12.142148448 +0000 UTC m=+830.332677094" watchObservedRunningTime="2025-10-09 08:31:12.143363793 +0000 UTC m=+830.333892419" Oct 09 08:31:19 crc kubenswrapper[4872]: I1009 08:31:19.166756 4872 generic.go:334] "Generic (PLEG): container finished" podID="94c5800f-97ae-4482-a959-d459cbdfdbaf" containerID="6eb9db79385ad4fc1faf8afd3a01ef363dfb9a2c73aea0cf461c47779fe00969" exitCode=0 Oct 09 08:31:19 crc kubenswrapper[4872]: I1009 08:31:19.166899 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerDied","Data":"6eb9db79385ad4fc1faf8afd3a01ef363dfb9a2c73aea0cf461c47779fe00969"} Oct 09 08:31:19 crc kubenswrapper[4872]: I1009 08:31:19.169471 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" event={"ID":"381ccfd5-05cf-4592-933e-128d714bf0a6","Type":"ContainerStarted","Data":"f4a9b5ba9133a273ab2f1280554af1f0a0f750bc7fe78ab3d2d4a26ece661b52"} Oct 09 08:31:19 crc kubenswrapper[4872]: I1009 08:31:19.169685 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:19 crc kubenswrapper[4872]: I1009 08:31:19.212281 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" podStartSLOduration=3.064605097 podStartE2EDuration="10.212265523s" podCreationTimestamp="2025-10-09 08:31:09 +0000 UTC" firstStartedPulling="2025-10-09 08:31:10.975243232 +0000 UTC m=+829.165771878" lastFinishedPulling="2025-10-09 08:31:18.122903668 +0000 UTC m=+836.313432304" observedRunningTime="2025-10-09 08:31:19.209340439 +0000 UTC m=+837.399869115" watchObservedRunningTime="2025-10-09 08:31:19.212265523 +0000 UTC m=+837.402794149" Oct 09 08:31:20 crc kubenswrapper[4872]: I1009 08:31:20.179418 4872 generic.go:334] "Generic (PLEG): container finished" podID="94c5800f-97ae-4482-a959-d459cbdfdbaf" containerID="0fde0eed92612416c1003168e433481bd77e29ba1dbd452e5af40f53530d1495" exitCode=0 Oct 09 08:31:20 crc kubenswrapper[4872]: I1009 08:31:20.179474 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerDied","Data":"0fde0eed92612416c1003168e433481bd77e29ba1dbd452e5af40f53530d1495"} Oct 09 08:31:21 crc kubenswrapper[4872]: I1009 08:31:21.190943 4872 generic.go:334] "Generic (PLEG): container finished" podID="94c5800f-97ae-4482-a959-d459cbdfdbaf" containerID="bd5d2c2af8d414c97a5adccbf80b783c4e6f2e181bbe46980860b65e9dcb9b1e" exitCode=0 Oct 09 08:31:21 crc kubenswrapper[4872]: I1009 08:31:21.191000 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerDied","Data":"bd5d2c2af8d414c97a5adccbf80b783c4e6f2e181bbe46980860b65e9dcb9b1e"} Oct 09 08:31:21 crc kubenswrapper[4872]: I1009 08:31:21.495269 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-b52wr" Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203359 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"ef1f5aef14a2ecf4dbac91da398ec4f6a1dbefbfa3b1db29f900323075194399"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203702 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"9ae68775a5e302c68ce9513743f807751a30008ff45edbff315c913b954fcd66"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203718 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"2e8539f8fd0dae48b2d660c0f8b1d3b9340c9d1967771d7cf44cab750a4e2c13"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203729 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"900f69b84c557763640d01c0c92949bf94b5a07e2cc833ccbbd480ca294e1bcb"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203739 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"4547fe7b3051c023469950cdbde86c41d4153f9d8b496f21a26cd7240612f0d0"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203750 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hd8n4" event={"ID":"94c5800f-97ae-4482-a959-d459cbdfdbaf","Type":"ContainerStarted","Data":"992cfb6235df001d9d20603da1a94ced3fd5c68437d95e42d1b11998ef59d388"} Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.203909 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:22 crc kubenswrapper[4872]: I1009 08:31:22.226901 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hd8n4" podStartSLOduration=5.717197023 podStartE2EDuration="13.226885514s" podCreationTimestamp="2025-10-09 08:31:09 +0000 UTC" firstStartedPulling="2025-10-09 08:31:10.635916284 +0000 UTC m=+828.826444910" lastFinishedPulling="2025-10-09 08:31:18.145604765 +0000 UTC m=+836.336133401" observedRunningTime="2025-10-09 08:31:22.22605196 +0000 UTC m=+840.416580636" watchObservedRunningTime="2025-10-09 08:31:22.226885514 +0000 UTC m=+840.417414140" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.498327 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.499267 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.501391 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.502075 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.510418 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-gp87t" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.530695 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.675924 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6t4h\" (UniqueName: \"kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h\") pod \"openstack-operator-index-kq4b6\" (UID: \"454cb5f5-4995-425d-807c-2194e893cc68\") " pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.777568 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6t4h\" (UniqueName: \"kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h\") pod \"openstack-operator-index-kq4b6\" (UID: \"454cb5f5-4995-425d-807c-2194e893cc68\") " pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.800285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6t4h\" (UniqueName: \"kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h\") pod \"openstack-operator-index-kq4b6\" (UID: \"454cb5f5-4995-425d-807c-2194e893cc68\") " pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:24 crc kubenswrapper[4872]: I1009 08:31:24.824316 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:25 crc kubenswrapper[4872]: I1009 08:31:25.207363 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:25 crc kubenswrapper[4872]: I1009 08:31:25.228079 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kq4b6" event={"ID":"454cb5f5-4995-425d-807c-2194e893cc68","Type":"ContainerStarted","Data":"a66e0d105ea94d68baa25ad6d4260887f79bc70c09653f2cd9e51958734d6961"} Oct 09 08:31:25 crc kubenswrapper[4872]: I1009 08:31:25.478603 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:25 crc kubenswrapper[4872]: I1009 08:31:25.530427 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:27 crc kubenswrapper[4872]: I1009 08:31:27.880020 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.249462 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kq4b6" event={"ID":"454cb5f5-4995-425d-807c-2194e893cc68","Type":"ContainerStarted","Data":"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9"} Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.280245 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-kq4b6" podStartSLOduration=1.743875176 podStartE2EDuration="4.280221923s" podCreationTimestamp="2025-10-09 08:31:24 +0000 UTC" firstStartedPulling="2025-10-09 08:31:25.221419036 +0000 UTC m=+843.411947662" lastFinishedPulling="2025-10-09 08:31:27.757765773 +0000 UTC m=+845.948294409" observedRunningTime="2025-10-09 08:31:28.272379737 +0000 UTC m=+846.462908473" watchObservedRunningTime="2025-10-09 08:31:28.280221923 +0000 UTC m=+846.470750559" Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.488498 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5jngg"] Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.489198 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.500754 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5jngg"] Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.634548 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbp7n\" (UniqueName: \"kubernetes.io/projected/6249ff92-0786-431e-9f18-d48d3b2a6863-kube-api-access-vbp7n\") pod \"openstack-operator-index-5jngg\" (UID: \"6249ff92-0786-431e-9f18-d48d3b2a6863\") " pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.738956 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbp7n\" (UniqueName: \"kubernetes.io/projected/6249ff92-0786-431e-9f18-d48d3b2a6863-kube-api-access-vbp7n\") pod \"openstack-operator-index-5jngg\" (UID: \"6249ff92-0786-431e-9f18-d48d3b2a6863\") " pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.773790 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbp7n\" (UniqueName: \"kubernetes.io/projected/6249ff92-0786-431e-9f18-d48d3b2a6863-kube-api-access-vbp7n\") pod \"openstack-operator-index-5jngg\" (UID: \"6249ff92-0786-431e-9f18-d48d3b2a6863\") " pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:28 crc kubenswrapper[4872]: I1009 08:31:28.846061 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.257464 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-kq4b6" podUID="454cb5f5-4995-425d-807c-2194e893cc68" containerName="registry-server" containerID="cri-o://3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9" gracePeriod=2 Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.287284 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5jngg"] Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.603879 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.728944 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-2hm62" Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.762766 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6t4h\" (UniqueName: \"kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h\") pod \"454cb5f5-4995-425d-807c-2194e893cc68\" (UID: \"454cb5f5-4995-425d-807c-2194e893cc68\") " Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.767268 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h" (OuterVolumeSpecName: "kube-api-access-c6t4h") pod "454cb5f5-4995-425d-807c-2194e893cc68" (UID: "454cb5f5-4995-425d-807c-2194e893cc68"). InnerVolumeSpecName "kube-api-access-c6t4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:31:29 crc kubenswrapper[4872]: I1009 08:31:29.865280 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6t4h\" (UniqueName: \"kubernetes.io/projected/454cb5f5-4995-425d-807c-2194e893cc68-kube-api-access-c6t4h\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.149560 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-9zqhg" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.266592 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-kq4b6" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.266730 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kq4b6" event={"ID":"454cb5f5-4995-425d-807c-2194e893cc68","Type":"ContainerDied","Data":"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9"} Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.266814 4872 scope.go:117] "RemoveContainer" containerID="3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.266608 4872 generic.go:334] "Generic (PLEG): container finished" podID="454cb5f5-4995-425d-807c-2194e893cc68" containerID="3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9" exitCode=0 Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.267938 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-kq4b6" event={"ID":"454cb5f5-4995-425d-807c-2194e893cc68","Type":"ContainerDied","Data":"a66e0d105ea94d68baa25ad6d4260887f79bc70c09653f2cd9e51958734d6961"} Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.270090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5jngg" event={"ID":"6249ff92-0786-431e-9f18-d48d3b2a6863","Type":"ContainerStarted","Data":"4ea00f696d182e51cfea21659665f71544dd760dfe41ff490768fc1473a7e4e9"} Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.270129 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5jngg" event={"ID":"6249ff92-0786-431e-9f18-d48d3b2a6863","Type":"ContainerStarted","Data":"ced5d6e5aab387c0c3fd213cea11e76fedf661bddea4a7399b137ee63685cdfe"} Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.286687 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5jngg" podStartSLOduration=2.228815243 podStartE2EDuration="2.286665915s" podCreationTimestamp="2025-10-09 08:31:28 +0000 UTC" firstStartedPulling="2025-10-09 08:31:29.297770883 +0000 UTC m=+847.488299509" lastFinishedPulling="2025-10-09 08:31:29.355621555 +0000 UTC m=+847.546150181" observedRunningTime="2025-10-09 08:31:30.283187894 +0000 UTC m=+848.473716560" watchObservedRunningTime="2025-10-09 08:31:30.286665915 +0000 UTC m=+848.477194551" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.287228 4872 scope.go:117] "RemoveContainer" containerID="3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9" Oct 09 08:31:30 crc kubenswrapper[4872]: E1009 08:31:30.287607 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9\": container with ID starting with 3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9 not found: ID does not exist" containerID="3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.287670 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9"} err="failed to get container status \"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9\": rpc error: code = NotFound desc = could not find container \"3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9\": container with ID starting with 3afe547147954388a06b3fbf14d8ece927e301a80d11f458982ece4ef4cd79c9 not found: ID does not exist" Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.301944 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.304794 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-kq4b6"] Oct 09 08:31:30 crc kubenswrapper[4872]: I1009 08:31:30.469113 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="454cb5f5-4995-425d-807c-2194e893cc68" path="/var/lib/kubelet/pods/454cb5f5-4995-425d-807c-2194e893cc68/volumes" Oct 09 08:31:38 crc kubenswrapper[4872]: I1009 08:31:38.847457 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:38 crc kubenswrapper[4872]: I1009 08:31:38.848052 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:38 crc kubenswrapper[4872]: I1009 08:31:38.882918 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:39 crc kubenswrapper[4872]: I1009 08:31:39.376514 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5jngg" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.483580 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hd8n4" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.498248 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tj85s"] Oct 09 08:31:40 crc kubenswrapper[4872]: E1009 08:31:40.498615 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="454cb5f5-4995-425d-807c-2194e893cc68" containerName="registry-server" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.498667 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="454cb5f5-4995-425d-807c-2194e893cc68" containerName="registry-server" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.498886 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="454cb5f5-4995-425d-807c-2194e893cc68" containerName="registry-server" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.500500 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.510852 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tj85s"] Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.623433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-utilities\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.623731 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msf6s\" (UniqueName: \"kubernetes.io/projected/ee093ff4-27c6-46de-abb9-79df201ca9ea-kube-api-access-msf6s\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.623764 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-catalog-content\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.726785 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-utilities\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.726910 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msf6s\" (UniqueName: \"kubernetes.io/projected/ee093ff4-27c6-46de-abb9-79df201ca9ea-kube-api-access-msf6s\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.726972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-catalog-content\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.727513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-utilities\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.727874 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee093ff4-27c6-46de-abb9-79df201ca9ea-catalog-content\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.753065 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msf6s\" (UniqueName: \"kubernetes.io/projected/ee093ff4-27c6-46de-abb9-79df201ca9ea-kube-api-access-msf6s\") pod \"certified-operators-tj85s\" (UID: \"ee093ff4-27c6-46de-abb9-79df201ca9ea\") " pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:40 crc kubenswrapper[4872]: I1009 08:31:40.863004 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:41 crc kubenswrapper[4872]: I1009 08:31:41.306576 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tj85s"] Oct 09 08:31:41 crc kubenswrapper[4872]: I1009 08:31:41.350589 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tj85s" event={"ID":"ee093ff4-27c6-46de-abb9-79df201ca9ea","Type":"ContainerStarted","Data":"12e663ea82931edbfcf7949b7a53d370f74a5d5462308364e6edf222ebf53969"} Oct 09 08:31:42 crc kubenswrapper[4872]: I1009 08:31:42.360072 4872 generic.go:334] "Generic (PLEG): container finished" podID="ee093ff4-27c6-46de-abb9-79df201ca9ea" containerID="634105af8d3eb176863971540635438dbbc8b802f23b03eb4b113b892f591ad0" exitCode=0 Oct 09 08:31:42 crc kubenswrapper[4872]: I1009 08:31:42.360129 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tj85s" event={"ID":"ee093ff4-27c6-46de-abb9-79df201ca9ea","Type":"ContainerDied","Data":"634105af8d3eb176863971540635438dbbc8b802f23b03eb4b113b892f591ad0"} Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.699735 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.701139 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.705176 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.768495 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.768593 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.768655 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g7wv\" (UniqueName: \"kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.869568 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.869625 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.869681 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g7wv\" (UniqueName: \"kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.870197 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.870258 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:43 crc kubenswrapper[4872]: I1009 08:31:43.902000 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g7wv\" (UniqueName: \"kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv\") pod \"redhat-marketplace-qfkjz\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:44 crc kubenswrapper[4872]: I1009 08:31:44.038094 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:44 crc kubenswrapper[4872]: I1009 08:31:44.458556 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:44 crc kubenswrapper[4872]: W1009 08:31:44.468693 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda026abc2_1e37_49be_9927_ee94411ff6e7.slice/crio-23a296e4b6fd74200767f9c5142320024ccb49fe8cf8ad4aecb7f1a22aa47f73 WatchSource:0}: Error finding container 23a296e4b6fd74200767f9c5142320024ccb49fe8cf8ad4aecb7f1a22aa47f73: Status 404 returned error can't find the container with id 23a296e4b6fd74200767f9c5142320024ccb49fe8cf8ad4aecb7f1a22aa47f73 Oct 09 08:31:45 crc kubenswrapper[4872]: I1009 08:31:45.376461 4872 generic.go:334] "Generic (PLEG): container finished" podID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerID="662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3" exitCode=0 Oct 09 08:31:45 crc kubenswrapper[4872]: I1009 08:31:45.376502 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerDied","Data":"662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3"} Oct 09 08:31:45 crc kubenswrapper[4872]: I1009 08:31:45.376529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerStarted","Data":"23a296e4b6fd74200767f9c5142320024ccb49fe8cf8ad4aecb7f1a22aa47f73"} Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.387549 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tj85s" event={"ID":"ee093ff4-27c6-46de-abb9-79df201ca9ea","Type":"ContainerStarted","Data":"e16d8272f6535c938a7546f45faa59e654d2f094b67d88caba4da3184771e9f8"} Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.533276 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs"] Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.534977 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.537381 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wwz2g" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.551191 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs"] Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.612809 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.612871 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjvsz\" (UniqueName: \"kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.612900 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.714117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.714201 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjvsz\" (UniqueName: \"kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.714246 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.714791 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.714899 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.738356 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjvsz\" (UniqueName: \"kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz\") pod \"01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:46 crc kubenswrapper[4872]: I1009 08:31:46.859304 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.298614 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs"] Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.395346 4872 generic.go:334] "Generic (PLEG): container finished" podID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerID="6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992" exitCode=0 Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.395450 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerDied","Data":"6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992"} Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.396433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" event={"ID":"6e67af22-d4fc-40de-b757-d08690f67c9c","Type":"ContainerStarted","Data":"9f3715ce0e6fa7a438f399bf7cd0263f39c7cb1e2107179099a89edb4d8ace77"} Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.403751 4872 generic.go:334] "Generic (PLEG): container finished" podID="ee093ff4-27c6-46de-abb9-79df201ca9ea" containerID="e16d8272f6535c938a7546f45faa59e654d2f094b67d88caba4da3184771e9f8" exitCode=0 Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.403787 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tj85s" event={"ID":"ee093ff4-27c6-46de-abb9-79df201ca9ea","Type":"ContainerDied","Data":"e16d8272f6535c938a7546f45faa59e654d2f094b67d88caba4da3184771e9f8"} Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.403811 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tj85s" event={"ID":"ee093ff4-27c6-46de-abb9-79df201ca9ea","Type":"ContainerStarted","Data":"a8bf872bcadb198c8a314d9c7eef6e6b4abcf9080260bad533aae2b899e2ffdc"} Oct 09 08:31:47 crc kubenswrapper[4872]: I1009 08:31:47.441679 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tj85s" podStartSLOduration=3.030399903 podStartE2EDuration="7.441660651s" podCreationTimestamp="2025-10-09 08:31:40 +0000 UTC" firstStartedPulling="2025-10-09 08:31:42.361940574 +0000 UTC m=+860.552469200" lastFinishedPulling="2025-10-09 08:31:46.773201312 +0000 UTC m=+864.963729948" observedRunningTime="2025-10-09 08:31:47.440055945 +0000 UTC m=+865.630584581" watchObservedRunningTime="2025-10-09 08:31:47.441660651 +0000 UTC m=+865.632189287" Oct 09 08:31:48 crc kubenswrapper[4872]: I1009 08:31:48.412534 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerStarted","Data":"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f"} Oct 09 08:31:48 crc kubenswrapper[4872]: I1009 08:31:48.414145 4872 generic.go:334] "Generic (PLEG): container finished" podID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerID="7ca10178d0187cfc621e9fb3a14d8985578fdd2540292b5f72ad9f1073dccfe0" exitCode=0 Oct 09 08:31:48 crc kubenswrapper[4872]: I1009 08:31:48.414233 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" event={"ID":"6e67af22-d4fc-40de-b757-d08690f67c9c","Type":"ContainerDied","Data":"7ca10178d0187cfc621e9fb3a14d8985578fdd2540292b5f72ad9f1073dccfe0"} Oct 09 08:31:48 crc kubenswrapper[4872]: I1009 08:31:48.449265 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qfkjz" podStartSLOduration=3.410771356 podStartE2EDuration="5.449239003s" podCreationTimestamp="2025-10-09 08:31:43 +0000 UTC" firstStartedPulling="2025-10-09 08:31:45.882309472 +0000 UTC m=+864.072838098" lastFinishedPulling="2025-10-09 08:31:47.920777119 +0000 UTC m=+866.111305745" observedRunningTime="2025-10-09 08:31:48.44602728 +0000 UTC m=+866.636555906" watchObservedRunningTime="2025-10-09 08:31:48.449239003 +0000 UTC m=+866.639767629" Oct 09 08:31:49 crc kubenswrapper[4872]: I1009 08:31:49.422711 4872 generic.go:334] "Generic (PLEG): container finished" podID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerID="40b0744bb54f2308c4a21d97cd273eba8849dddad75c674995bb342d3057be45" exitCode=0 Oct 09 08:31:49 crc kubenswrapper[4872]: I1009 08:31:49.422821 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" event={"ID":"6e67af22-d4fc-40de-b757-d08690f67c9c","Type":"ContainerDied","Data":"40b0744bb54f2308c4a21d97cd273eba8849dddad75c674995bb342d3057be45"} Oct 09 08:31:50 crc kubenswrapper[4872]: I1009 08:31:50.429186 4872 generic.go:334] "Generic (PLEG): container finished" podID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerID="99ce230fa0ddfa61e015bcfc9143a0623dc56da9b208e958a388e172ba9baa3f" exitCode=0 Oct 09 08:31:50 crc kubenswrapper[4872]: I1009 08:31:50.429228 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" event={"ID":"6e67af22-d4fc-40de-b757-d08690f67c9c","Type":"ContainerDied","Data":"99ce230fa0ddfa61e015bcfc9143a0623dc56da9b208e958a388e172ba9baa3f"} Oct 09 08:31:50 crc kubenswrapper[4872]: I1009 08:31:50.863657 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:50 crc kubenswrapper[4872]: I1009 08:31:50.863723 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:50 crc kubenswrapper[4872]: I1009 08:31:50.940528 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.482347 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tj85s" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.693976 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.876812 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjvsz\" (UniqueName: \"kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz\") pod \"6e67af22-d4fc-40de-b757-d08690f67c9c\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.877080 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util\") pod \"6e67af22-d4fc-40de-b757-d08690f67c9c\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.877196 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle\") pod \"6e67af22-d4fc-40de-b757-d08690f67c9c\" (UID: \"6e67af22-d4fc-40de-b757-d08690f67c9c\") " Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.878294 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle" (OuterVolumeSpecName: "bundle") pod "6e67af22-d4fc-40de-b757-d08690f67c9c" (UID: "6e67af22-d4fc-40de-b757-d08690f67c9c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.885485 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz" (OuterVolumeSpecName: "kube-api-access-zjvsz") pod "6e67af22-d4fc-40de-b757-d08690f67c9c" (UID: "6e67af22-d4fc-40de-b757-d08690f67c9c"). InnerVolumeSpecName "kube-api-access-zjvsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.891442 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util" (OuterVolumeSpecName: "util") pod "6e67af22-d4fc-40de-b757-d08690f67c9c" (UID: "6e67af22-d4fc-40de-b757-d08690f67c9c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.979403 4872 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.979441 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjvsz\" (UniqueName: \"kubernetes.io/projected/6e67af22-d4fc-40de-b757-d08690f67c9c-kube-api-access-zjvsz\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:51 crc kubenswrapper[4872]: I1009 08:31:51.979452 4872 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/6e67af22-d4fc-40de-b757-d08690f67c9c-util\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:52 crc kubenswrapper[4872]: I1009 08:31:52.442834 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" Oct 09 08:31:52 crc kubenswrapper[4872]: I1009 08:31:52.442826 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs" event={"ID":"6e67af22-d4fc-40de-b757-d08690f67c9c","Type":"ContainerDied","Data":"9f3715ce0e6fa7a438f399bf7cd0263f39c7cb1e2107179099a89edb4d8ace77"} Oct 09 08:31:52 crc kubenswrapper[4872]: I1009 08:31:52.443006 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f3715ce0e6fa7a438f399bf7cd0263f39c7cb1e2107179099a89edb4d8ace77" Oct 09 08:31:54 crc kubenswrapper[4872]: I1009 08:31:54.038990 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:54 crc kubenswrapper[4872]: I1009 08:31:54.039354 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:54 crc kubenswrapper[4872]: I1009 08:31:54.116901 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:54 crc kubenswrapper[4872]: I1009 08:31:54.496684 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:55 crc kubenswrapper[4872]: I1009 08:31:55.113526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tj85s"] Oct 09 08:31:55 crc kubenswrapper[4872]: I1009 08:31:55.680672 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:31:55 crc kubenswrapper[4872]: I1009 08:31:55.680962 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mxtxg" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="registry-server" containerID="cri-o://034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828" gracePeriod=2 Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.082563 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.134133 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities\") pod \"76a38bf4-890c-4f97-a34d-33cc8adbc882\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.134173 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content\") pod \"76a38bf4-890c-4f97-a34d-33cc8adbc882\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.134209 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ck5w\" (UniqueName: \"kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w\") pod \"76a38bf4-890c-4f97-a34d-33cc8adbc882\" (UID: \"76a38bf4-890c-4f97-a34d-33cc8adbc882\") " Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.135047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities" (OuterVolumeSpecName: "utilities") pod "76a38bf4-890c-4f97-a34d-33cc8adbc882" (UID: "76a38bf4-890c-4f97-a34d-33cc8adbc882"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.141869 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w" (OuterVolumeSpecName: "kube-api-access-8ck5w") pod "76a38bf4-890c-4f97-a34d-33cc8adbc882" (UID: "76a38bf4-890c-4f97-a34d-33cc8adbc882"). InnerVolumeSpecName "kube-api-access-8ck5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.187625 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76a38bf4-890c-4f97-a34d-33cc8adbc882" (UID: "76a38bf4-890c-4f97-a34d-33cc8adbc882"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.235437 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ck5w\" (UniqueName: \"kubernetes.io/projected/76a38bf4-890c-4f97-a34d-33cc8adbc882-kube-api-access-8ck5w\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.235477 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.235488 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a38bf4-890c-4f97-a34d-33cc8adbc882-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.470347 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerID="034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828" exitCode=0 Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.470392 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerDied","Data":"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828"} Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.470419 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxtxg" event={"ID":"76a38bf4-890c-4f97-a34d-33cc8adbc882","Type":"ContainerDied","Data":"094e230a91ee796a38f8b920ec9ab99bacf23906fbf2fa79ded8f87e45199fbc"} Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.470439 4872 scope.go:117] "RemoveContainer" containerID="034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.470452 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxtxg" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.496040 4872 scope.go:117] "RemoveContainer" containerID="68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.504613 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.510110 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mxtxg"] Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.530019 4872 scope.go:117] "RemoveContainer" containerID="8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.550355 4872 scope.go:117] "RemoveContainer" containerID="034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828" Oct 09 08:31:56 crc kubenswrapper[4872]: E1009 08:31:56.551437 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828\": container with ID starting with 034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828 not found: ID does not exist" containerID="034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.553011 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828"} err="failed to get container status \"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828\": rpc error: code = NotFound desc = could not find container \"034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828\": container with ID starting with 034d8a5fa405d0301d1bed4b204c987c26f01facea95c2a99a6e071636ba0828 not found: ID does not exist" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.553159 4872 scope.go:117] "RemoveContainer" containerID="68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b" Oct 09 08:31:56 crc kubenswrapper[4872]: E1009 08:31:56.553918 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b\": container with ID starting with 68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b not found: ID does not exist" containerID="68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.554059 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b"} err="failed to get container status \"68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b\": rpc error: code = NotFound desc = could not find container \"68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b\": container with ID starting with 68842d983b6b8a357e52fb9324102855df9f79a99965684b4fd172999348ae3b not found: ID does not exist" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.554637 4872 scope.go:117] "RemoveContainer" containerID="8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1" Oct 09 08:31:56 crc kubenswrapper[4872]: E1009 08:31:56.555712 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1\": container with ID starting with 8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1 not found: ID does not exist" containerID="8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1" Oct 09 08:31:56 crc kubenswrapper[4872]: I1009 08:31:56.555778 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1"} err="failed to get container status \"8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1\": rpc error: code = NotFound desc = could not find container \"8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1\": container with ID starting with 8b3814f709c0a9c1433005318fb808a6fce27b280f31d791425c0a1f3ef2a5a1 not found: ID does not exist" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.664702 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9"] Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.665694 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="extract" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.665785 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="extract" Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.665851 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="extract-utilities" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.665910 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="extract-utilities" Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.665970 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="extract-content" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666022 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="extract-content" Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.666097 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="util" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666158 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="util" Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.666226 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="pull" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666288 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="pull" Oct 09 08:31:57 crc kubenswrapper[4872]: E1009 08:31:57.666354 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="registry-server" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666415 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="registry-server" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666620 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" containerName="registry-server" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.666724 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e67af22-d4fc-40de-b757-d08690f67c9c" containerName="extract" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.667480 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.677045 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-8gg2l" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.679967 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.680211 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qfkjz" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="registry-server" containerID="cri-o://44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f" gracePeriod=2 Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.693264 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9"] Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.755357 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nh8l7\" (UniqueName: \"kubernetes.io/projected/02685dc1-985a-4360-8192-74858bb6062b-kube-api-access-nh8l7\") pod \"openstack-operator-controller-operator-557bb754c5-pj5t9\" (UID: \"02685dc1-985a-4360-8192-74858bb6062b\") " pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.863075 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nh8l7\" (UniqueName: \"kubernetes.io/projected/02685dc1-985a-4360-8192-74858bb6062b-kube-api-access-nh8l7\") pod \"openstack-operator-controller-operator-557bb754c5-pj5t9\" (UID: \"02685dc1-985a-4360-8192-74858bb6062b\") " pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.890780 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nh8l7\" (UniqueName: \"kubernetes.io/projected/02685dc1-985a-4360-8192-74858bb6062b-kube-api-access-nh8l7\") pod \"openstack-operator-controller-operator-557bb754c5-pj5t9\" (UID: \"02685dc1-985a-4360-8192-74858bb6062b\") " pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:31:57 crc kubenswrapper[4872]: I1009 08:31:57.984057 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.037560 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.071200 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content\") pod \"a026abc2-1e37-49be-9927-ee94411ff6e7\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.071251 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities\") pod \"a026abc2-1e37-49be-9927-ee94411ff6e7\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.071318 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g7wv\" (UniqueName: \"kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv\") pod \"a026abc2-1e37-49be-9927-ee94411ff6e7\" (UID: \"a026abc2-1e37-49be-9927-ee94411ff6e7\") " Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.075222 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv" (OuterVolumeSpecName: "kube-api-access-6g7wv") pod "a026abc2-1e37-49be-9927-ee94411ff6e7" (UID: "a026abc2-1e37-49be-9927-ee94411ff6e7"). InnerVolumeSpecName "kube-api-access-6g7wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.075230 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities" (OuterVolumeSpecName: "utilities") pod "a026abc2-1e37-49be-9927-ee94411ff6e7" (UID: "a026abc2-1e37-49be-9927-ee94411ff6e7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.100938 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a026abc2-1e37-49be-9927-ee94411ff6e7" (UID: "a026abc2-1e37-49be-9927-ee94411ff6e7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.173055 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.173111 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a026abc2-1e37-49be-9927-ee94411ff6e7-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.173125 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g7wv\" (UniqueName: \"kubernetes.io/projected/a026abc2-1e37-49be-9927-ee94411ff6e7-kube-api-access-6g7wv\") on node \"crc\" DevicePath \"\"" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.412715 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9"] Oct 09 08:31:58 crc kubenswrapper[4872]: W1009 08:31:58.420375 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod02685dc1_985a_4360_8192_74858bb6062b.slice/crio-6796b2bcf53f8ef70281b8debb9b40eb5686a3aef60ddd748ac04cd200862a60 WatchSource:0}: Error finding container 6796b2bcf53f8ef70281b8debb9b40eb5686a3aef60ddd748ac04cd200862a60: Status 404 returned error can't find the container with id 6796b2bcf53f8ef70281b8debb9b40eb5686a3aef60ddd748ac04cd200862a60 Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.468786 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a38bf4-890c-4f97-a34d-33cc8adbc882" path="/var/lib/kubelet/pods/76a38bf4-890c-4f97-a34d-33cc8adbc882/volumes" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.497552 4872 generic.go:334] "Generic (PLEG): container finished" podID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerID="44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f" exitCode=0 Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.497670 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerDied","Data":"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f"} Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.497709 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfkjz" event={"ID":"a026abc2-1e37-49be-9927-ee94411ff6e7","Type":"ContainerDied","Data":"23a296e4b6fd74200767f9c5142320024ccb49fe8cf8ad4aecb7f1a22aa47f73"} Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.497727 4872 scope.go:117] "RemoveContainer" containerID="44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.497870 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfkjz" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.502895 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" event={"ID":"02685dc1-985a-4360-8192-74858bb6062b","Type":"ContainerStarted","Data":"6796b2bcf53f8ef70281b8debb9b40eb5686a3aef60ddd748ac04cd200862a60"} Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.527591 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.528321 4872 scope.go:117] "RemoveContainer" containerID="6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.530991 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfkjz"] Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.562496 4872 scope.go:117] "RemoveContainer" containerID="662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.580230 4872 scope.go:117] "RemoveContainer" containerID="44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f" Oct 09 08:31:58 crc kubenswrapper[4872]: E1009 08:31:58.580849 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f\": container with ID starting with 44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f not found: ID does not exist" containerID="44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.581116 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f"} err="failed to get container status \"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f\": rpc error: code = NotFound desc = could not find container \"44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f\": container with ID starting with 44f09cd6692c2ee4c9baf8359fed30d36e1da7468b3b1034817b94b817a75a6f not found: ID does not exist" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.581179 4872 scope.go:117] "RemoveContainer" containerID="6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992" Oct 09 08:31:58 crc kubenswrapper[4872]: E1009 08:31:58.583660 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992\": container with ID starting with 6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992 not found: ID does not exist" containerID="6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.583772 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992"} err="failed to get container status \"6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992\": rpc error: code = NotFound desc = could not find container \"6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992\": container with ID starting with 6fc10eb1459a79e16dbd92a8324298747e939d91a697880f5488e16e9ac8f992 not found: ID does not exist" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.583796 4872 scope.go:117] "RemoveContainer" containerID="662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3" Oct 09 08:31:58 crc kubenswrapper[4872]: E1009 08:31:58.584035 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3\": container with ID starting with 662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3 not found: ID does not exist" containerID="662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3" Oct 09 08:31:58 crc kubenswrapper[4872]: I1009 08:31:58.585841 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3"} err="failed to get container status \"662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3\": rpc error: code = NotFound desc = could not find container \"662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3\": container with ID starting with 662761643b2a4d9fe3fe30c8eb727f424d7b2e749e125e5f6f15637218a9f5c3 not found: ID does not exist" Oct 09 08:32:00 crc kubenswrapper[4872]: I1009 08:32:00.471294 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" path="/var/lib/kubelet/pods/a026abc2-1e37-49be-9927-ee94411ff6e7/volumes" Oct 09 08:32:02 crc kubenswrapper[4872]: I1009 08:32:02.531878 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" event={"ID":"02685dc1-985a-4360-8192-74858bb6062b","Type":"ContainerStarted","Data":"da7a7cdfbfa26d03360dfc16daf235f1f3559b10d260154d325673022a2a599b"} Oct 09 08:32:04 crc kubenswrapper[4872]: I1009 08:32:04.547749 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" event={"ID":"02685dc1-985a-4360-8192-74858bb6062b","Type":"ContainerStarted","Data":"cb61e87c00df79711a07f30ba1e157219e77078f448bd589234887984c68ef17"} Oct 09 08:32:04 crc kubenswrapper[4872]: I1009 08:32:04.548396 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:32:04 crc kubenswrapper[4872]: I1009 08:32:04.579182 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" podStartSLOduration=1.5914342879999999 podStartE2EDuration="7.579161481s" podCreationTimestamp="2025-10-09 08:31:57 +0000 UTC" firstStartedPulling="2025-10-09 08:31:58.422401833 +0000 UTC m=+876.612930469" lastFinishedPulling="2025-10-09 08:32:04.410129036 +0000 UTC m=+882.600657662" observedRunningTime="2025-10-09 08:32:04.574721963 +0000 UTC m=+882.765250599" watchObservedRunningTime="2025-10-09 08:32:04.579161481 +0000 UTC m=+882.769690107" Oct 09 08:32:07 crc kubenswrapper[4872]: I1009 08:32:07.987721 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-557bb754c5-pj5t9" Oct 09 08:32:08 crc kubenswrapper[4872]: I1009 08:32:08.157695 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:32:08 crc kubenswrapper[4872]: I1009 08:32:08.157773 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.414084 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:18 crc kubenswrapper[4872]: E1009 08:32:18.415335 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="extract-utilities" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.415359 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="extract-utilities" Oct 09 08:32:18 crc kubenswrapper[4872]: E1009 08:32:18.415429 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="registry-server" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.415442 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="registry-server" Oct 09 08:32:18 crc kubenswrapper[4872]: E1009 08:32:18.415533 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="extract-content" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.415553 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="extract-content" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.415808 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a026abc2-1e37-49be-9927-ee94411ff6e7" containerName="registry-server" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.418137 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.423825 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.460244 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fq2bp\" (UniqueName: \"kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.460377 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.460464 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.561876 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fq2bp\" (UniqueName: \"kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.562188 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.562348 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.562733 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.562979 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.583040 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fq2bp\" (UniqueName: \"kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp\") pod \"community-operators-b9vs7\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:18 crc kubenswrapper[4872]: I1009 08:32:18.736510 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:19 crc kubenswrapper[4872]: I1009 08:32:19.236284 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:19 crc kubenswrapper[4872]: I1009 08:32:19.654368 4872 generic.go:334] "Generic (PLEG): container finished" podID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerID="1aa01068d94adcf3dfdee60440dd67b929190cdd72d26825bba962f90e150cad" exitCode=0 Oct 09 08:32:19 crc kubenswrapper[4872]: I1009 08:32:19.654502 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerDied","Data":"1aa01068d94adcf3dfdee60440dd67b929190cdd72d26825bba962f90e150cad"} Oct 09 08:32:19 crc kubenswrapper[4872]: I1009 08:32:19.654544 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerStarted","Data":"f4506afaa0837e52769adf02a2d9d5fe25034f7a6b4e3a60df43fc630db58a80"} Oct 09 08:32:21 crc kubenswrapper[4872]: I1009 08:32:21.671412 4872 generic.go:334] "Generic (PLEG): container finished" podID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerID="d643ecb6f7efadf6ec132ae59d3778c9ee326de4d14936ec38fd9cabb085d77f" exitCode=0 Oct 09 08:32:21 crc kubenswrapper[4872]: I1009 08:32:21.671488 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerDied","Data":"d643ecb6f7efadf6ec132ae59d3778c9ee326de4d14936ec38fd9cabb085d77f"} Oct 09 08:32:22 crc kubenswrapper[4872]: I1009 08:32:22.681230 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerStarted","Data":"e453589fdf25fa6735419d8c1b921cbad5a16225f26d14d5f107394ff7a70891"} Oct 09 08:32:22 crc kubenswrapper[4872]: I1009 08:32:22.709829 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b9vs7" podStartSLOduration=1.972041157 podStartE2EDuration="4.709801576s" podCreationTimestamp="2025-10-09 08:32:18 +0000 UTC" firstStartedPulling="2025-10-09 08:32:19.656307321 +0000 UTC m=+897.846835997" lastFinishedPulling="2025-10-09 08:32:22.39406779 +0000 UTC m=+900.584596416" observedRunningTime="2025-10-09 08:32:22.702040292 +0000 UTC m=+900.892568958" watchObservedRunningTime="2025-10-09 08:32:22.709801576 +0000 UTC m=+900.900330232" Oct 09 08:32:28 crc kubenswrapper[4872]: I1009 08:32:28.737266 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:28 crc kubenswrapper[4872]: I1009 08:32:28.737917 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:28 crc kubenswrapper[4872]: I1009 08:32:28.778045 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:29 crc kubenswrapper[4872]: I1009 08:32:29.766692 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:29 crc kubenswrapper[4872]: I1009 08:32:29.801844 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:31 crc kubenswrapper[4872]: I1009 08:32:31.737119 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b9vs7" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="registry-server" containerID="cri-o://e453589fdf25fa6735419d8c1b921cbad5a16225f26d14d5f107394ff7a70891" gracePeriod=2 Oct 09 08:32:32 crc kubenswrapper[4872]: I1009 08:32:32.746537 4872 generic.go:334] "Generic (PLEG): container finished" podID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerID="e453589fdf25fa6735419d8c1b921cbad5a16225f26d14d5f107394ff7a70891" exitCode=0 Oct 09 08:32:32 crc kubenswrapper[4872]: I1009 08:32:32.746589 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerDied","Data":"e453589fdf25fa6735419d8c1b921cbad5a16225f26d14d5f107394ff7a70891"} Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.553930 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.680791 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content\") pod \"c789d2c2-e9cd-4a46-a83a-607769ac1558\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.680911 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fq2bp\" (UniqueName: \"kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp\") pod \"c789d2c2-e9cd-4a46-a83a-607769ac1558\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.681017 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities\") pod \"c789d2c2-e9cd-4a46-a83a-607769ac1558\" (UID: \"c789d2c2-e9cd-4a46-a83a-607769ac1558\") " Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.681950 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities" (OuterVolumeSpecName: "utilities") pod "c789d2c2-e9cd-4a46-a83a-607769ac1558" (UID: "c789d2c2-e9cd-4a46-a83a-607769ac1558"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.688211 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp" (OuterVolumeSpecName: "kube-api-access-fq2bp") pod "c789d2c2-e9cd-4a46-a83a-607769ac1558" (UID: "c789d2c2-e9cd-4a46-a83a-607769ac1558"). InnerVolumeSpecName "kube-api-access-fq2bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.730577 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c789d2c2-e9cd-4a46-a83a-607769ac1558" (UID: "c789d2c2-e9cd-4a46-a83a-607769ac1558"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.754293 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b9vs7" event={"ID":"c789d2c2-e9cd-4a46-a83a-607769ac1558","Type":"ContainerDied","Data":"f4506afaa0837e52769adf02a2d9d5fe25034f7a6b4e3a60df43fc630db58a80"} Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.754353 4872 scope.go:117] "RemoveContainer" containerID="e453589fdf25fa6735419d8c1b921cbad5a16225f26d14d5f107394ff7a70891" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.754380 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b9vs7" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.781209 4872 scope.go:117] "RemoveContainer" containerID="d643ecb6f7efadf6ec132ae59d3778c9ee326de4d14936ec38fd9cabb085d77f" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.781815 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.782659 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fq2bp\" (UniqueName: \"kubernetes.io/projected/c789d2c2-e9cd-4a46-a83a-607769ac1558-kube-api-access-fq2bp\") on node \"crc\" DevicePath \"\"" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.782678 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.782688 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c789d2c2-e9cd-4a46-a83a-607769ac1558-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.796776 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b9vs7"] Oct 09 08:32:33 crc kubenswrapper[4872]: I1009 08:32:33.806603 4872 scope.go:117] "RemoveContainer" containerID="1aa01068d94adcf3dfdee60440dd67b929190cdd72d26825bba962f90e150cad" Oct 09 08:32:34 crc kubenswrapper[4872]: I1009 08:32:34.468045 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" path="/var/lib/kubelet/pods/c789d2c2-e9cd-4a46-a83a-607769ac1558/volumes" Oct 09 08:32:38 crc kubenswrapper[4872]: I1009 08:32:38.157257 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:32:38 crc kubenswrapper[4872]: I1009 08:32:38.157611 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.292141 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc"] Oct 09 08:32:44 crc kubenswrapper[4872]: E1009 08:32:44.293032 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="extract-utilities" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.293050 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="extract-utilities" Oct 09 08:32:44 crc kubenswrapper[4872]: E1009 08:32:44.293082 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="registry-server" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.293091 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="registry-server" Oct 09 08:32:44 crc kubenswrapper[4872]: E1009 08:32:44.293111 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="extract-content" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.293120 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="extract-content" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.293261 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c789d2c2-e9cd-4a46-a83a-607769ac1558" containerName="registry-server" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.294028 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.297037 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-mwlbj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.328943 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.338873 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.339850 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.388326 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-4pqht" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.418767 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.429818 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.434737 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lfz8\" (UniqueName: \"kubernetes.io/projected/91db56d0-f7bc-4d02-92ec-c7ead6625176-kube-api-access-2lfz8\") pod \"barbican-operator-controller-manager-64f84fcdbb-lrlzc\" (UID: \"91db56d0-f7bc-4d02-92ec-c7ead6625176\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.434876 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4v8g4\" (UniqueName: \"kubernetes.io/projected/283f56cf-99e9-471b-8f40-10ec0f31ad06-kube-api-access-4v8g4\") pod \"cinder-operator-controller-manager-59cdc64769-vsdt6\" (UID: \"283f56cf-99e9-471b-8f40-10ec0f31ad06\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.480227 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.497996 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-gkjxq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.501106 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.502856 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.504027 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.504738 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.506670 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.507469 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.508420 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-6ntn7" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.508845 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-zt8pj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.512614 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-7xsh9" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.520205 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.531899 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.547683 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.551492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb95q\" (UniqueName: \"kubernetes.io/projected/b9b5d69d-fb26-4286-a9b0-3b00d042972a-kube-api-access-mb95q\") pod \"heat-operator-controller-manager-6d9967f8dd-qnnp4\" (UID: \"b9b5d69d-fb26-4286-a9b0-3b00d042972a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.551592 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lfz8\" (UniqueName: \"kubernetes.io/projected/91db56d0-f7bc-4d02-92ec-c7ead6625176-kube-api-access-2lfz8\") pod \"barbican-operator-controller-manager-64f84fcdbb-lrlzc\" (UID: \"91db56d0-f7bc-4d02-92ec-c7ead6625176\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.551628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4v8g4\" (UniqueName: \"kubernetes.io/projected/283f56cf-99e9-471b-8f40-10ec0f31ad06-kube-api-access-4v8g4\") pod \"cinder-operator-controller-manager-59cdc64769-vsdt6\" (UID: \"283f56cf-99e9-471b-8f40-10ec0f31ad06\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.553992 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.555453 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.566773 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.567017 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-bvvzj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.569334 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.584128 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.584928 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.588847 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lfz8\" (UniqueName: \"kubernetes.io/projected/91db56d0-f7bc-4d02-92ec-c7ead6625176-kube-api-access-2lfz8\") pod \"barbican-operator-controller-manager-64f84fcdbb-lrlzc\" (UID: \"91db56d0-f7bc-4d02-92ec-c7ead6625176\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.590654 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-m6hcx" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.595499 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.596701 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.601191 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4v8g4\" (UniqueName: \"kubernetes.io/projected/283f56cf-99e9-471b-8f40-10ec0f31ad06-kube-api-access-4v8g4\") pod \"cinder-operator-controller-manager-59cdc64769-vsdt6\" (UID: \"283f56cf-99e9-471b-8f40-10ec0f31ad06\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.608375 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-hqtdl" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.608579 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.610615 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-2knwl"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.612144 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.620057 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.621016 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-4nxqs" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.642489 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656122 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-2knwl"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656163 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656248 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5mk\" (UniqueName: \"kubernetes.io/projected/fab5d9a4-352a-4092-9139-c74242142eda-kube-api-access-8d5mk\") pod \"glance-operator-controller-manager-7bb46cd7d-dskmk\" (UID: \"fab5d9a4-352a-4092-9139-c74242142eda\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656386 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgxv4\" (UniqueName: \"kubernetes.io/projected/57a99d6c-df10-4de9-9434-471ceb56878d-kube-api-access-pgxv4\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656417 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqn77\" (UniqueName: \"kubernetes.io/projected/c5d3f64a-0214-457c-b4c6-f48beabe8669-kube-api-access-vqn77\") pod \"ironic-operator-controller-manager-74cb5cbc49-vqj98\" (UID: \"c5d3f64a-0214-457c-b4c6-f48beabe8669\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656440 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8kn6\" (UniqueName: \"kubernetes.io/projected/0d04d185-02c6-4688-83cd-e9f1744776f1-kube-api-access-c8kn6\") pod \"designate-operator-controller-manager-687df44cdb-26dhq\" (UID: \"0d04d185-02c6-4688-83cd-e9f1744776f1\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clzcc\" (UniqueName: \"kubernetes.io/projected/a34a57a1-2b91-4d65-8336-4a363b33dd45-kube-api-access-clzcc\") pod \"keystone-operator-controller-manager-ddb98f99b-8rkjv\" (UID: \"a34a57a1-2b91-4d65-8336-4a363b33dd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656519 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb95q\" (UniqueName: \"kubernetes.io/projected/b9b5d69d-fb26-4286-a9b0-3b00d042972a-kube-api-access-mb95q\") pod \"heat-operator-controller-manager-6d9967f8dd-qnnp4\" (UID: \"b9b5d69d-fb26-4286-a9b0-3b00d042972a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656552 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hl8j4\" (UniqueName: \"kubernetes.io/projected/deb85a9b-515a-4027-a178-c83829fd5a34-kube-api-access-hl8j4\") pod \"horizon-operator-controller-manager-6d74794d9b-n5rbg\" (UID: \"deb85a9b-515a-4027-a178-c83829fd5a34\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.656571 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.661039 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-drks7" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.662664 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.669783 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.675671 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.676549 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.680559 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.681430 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.686228 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-8b57g" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.686510 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-5n5n4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.704464 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb95q\" (UniqueName: \"kubernetes.io/projected/b9b5d69d-fb26-4286-a9b0-3b00d042972a-kube-api-access-mb95q\") pod \"heat-operator-controller-manager-6d9967f8dd-qnnp4\" (UID: \"b9b5d69d-fb26-4286-a9b0-3b00d042972a\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.705301 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.714616 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.758206 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clzcc\" (UniqueName: \"kubernetes.io/projected/a34a57a1-2b91-4d65-8336-4a363b33dd45-kube-api-access-clzcc\") pod \"keystone-operator-controller-manager-ddb98f99b-8rkjv\" (UID: \"a34a57a1-2b91-4d65-8336-4a363b33dd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.758243 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n2j8\" (UniqueName: \"kubernetes.io/projected/18e70210-0a78-4ea0-8d8b-a955cdbec6b8-kube-api-access-6n2j8\") pod \"manila-operator-controller-manager-59578bc799-2knwl\" (UID: \"18e70210-0a78-4ea0-8d8b-a955cdbec6b8\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.758290 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hl8j4\" (UniqueName: \"kubernetes.io/projected/deb85a9b-515a-4027-a178-c83829fd5a34-kube-api-access-hl8j4\") pod \"horizon-operator-controller-manager-6d74794d9b-n5rbg\" (UID: \"deb85a9b-515a-4027-a178-c83829fd5a34\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.758311 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: E1009 08:32:44.758882 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 09 08:32:44 crc kubenswrapper[4872]: E1009 08:32:44.758922 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert podName:57a99d6c-df10-4de9-9434-471ceb56878d nodeName:}" failed. No retries permitted until 2025-10-09 08:32:45.258907249 +0000 UTC m=+923.449435875 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert") pod "infra-operator-controller-manager-585fc5b659-kgzkf" (UID: "57a99d6c-df10-4de9-9434-471ceb56878d") : secret "infra-operator-webhook-server-cert" not found Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759077 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnl4x\" (UniqueName: \"kubernetes.io/projected/ae4070fd-f49f-403d-a0b5-3285f235a221-kube-api-access-mnl4x\") pod \"nova-operator-controller-manager-57bb74c7bf-v67sz\" (UID: \"ae4070fd-f49f-403d-a0b5-3285f235a221\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759108 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5mk\" (UniqueName: \"kubernetes.io/projected/fab5d9a4-352a-4092-9139-c74242142eda-kube-api-access-8d5mk\") pod \"glance-operator-controller-manager-7bb46cd7d-dskmk\" (UID: \"fab5d9a4-352a-4092-9139-c74242142eda\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759133 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c54k\" (UniqueName: \"kubernetes.io/projected/88a9a65b-d7fe-4471-b408-562f5c98bdbb-kube-api-access-7c54k\") pod \"mariadb-operator-controller-manager-5777b4f897-8ttcw\" (UID: \"88a9a65b-d7fe-4471-b408-562f5c98bdbb\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759170 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgxv4\" (UniqueName: \"kubernetes.io/projected/57a99d6c-df10-4de9-9434-471ceb56878d-kube-api-access-pgxv4\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759192 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqn77\" (UniqueName: \"kubernetes.io/projected/c5d3f64a-0214-457c-b4c6-f48beabe8669-kube-api-access-vqn77\") pod \"ironic-operator-controller-manager-74cb5cbc49-vqj98\" (UID: \"c5d3f64a-0214-457c-b4c6-f48beabe8669\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759209 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8kn6\" (UniqueName: \"kubernetes.io/projected/0d04d185-02c6-4688-83cd-e9f1744776f1-kube-api-access-c8kn6\") pod \"designate-operator-controller-manager-687df44cdb-26dhq\" (UID: \"0d04d185-02c6-4688-83cd-e9f1744776f1\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.759233 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxdcn\" (UniqueName: \"kubernetes.io/projected/fcb8eabd-193f-4b03-b011-5e18514cc858-kube-api-access-xxdcn\") pod \"neutron-operator-controller-manager-797d478b46-kz58p\" (UID: \"fcb8eabd-193f-4b03-b011-5e18514cc858\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.764659 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.776621 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.777667 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.780715 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5h4r8" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.781805 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.782767 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.787682 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-5kr4b" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.791815 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.798181 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqn77\" (UniqueName: \"kubernetes.io/projected/c5d3f64a-0214-457c-b4c6-f48beabe8669-kube-api-access-vqn77\") pod \"ironic-operator-controller-manager-74cb5cbc49-vqj98\" (UID: \"c5d3f64a-0214-457c-b4c6-f48beabe8669\") " pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.802808 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5mk\" (UniqueName: \"kubernetes.io/projected/fab5d9a4-352a-4092-9139-c74242142eda-kube-api-access-8d5mk\") pod \"glance-operator-controller-manager-7bb46cd7d-dskmk\" (UID: \"fab5d9a4-352a-4092-9139-c74242142eda\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.803290 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clzcc\" (UniqueName: \"kubernetes.io/projected/a34a57a1-2b91-4d65-8336-4a363b33dd45-kube-api-access-clzcc\") pod \"keystone-operator-controller-manager-ddb98f99b-8rkjv\" (UID: \"a34a57a1-2b91-4d65-8336-4a363b33dd45\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.815985 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hl8j4\" (UniqueName: \"kubernetes.io/projected/deb85a9b-515a-4027-a178-c83829fd5a34-kube-api-access-hl8j4\") pod \"horizon-operator-controller-manager-6d74794d9b-n5rbg\" (UID: \"deb85a9b-515a-4027-a178-c83829fd5a34\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.817310 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgxv4\" (UniqueName: \"kubernetes.io/projected/57a99d6c-df10-4de9-9434-471ceb56878d-kube-api-access-pgxv4\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.825306 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.826235 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8kn6\" (UniqueName: \"kubernetes.io/projected/0d04d185-02c6-4688-83cd-e9f1744776f1-kube-api-access-c8kn6\") pod \"designate-operator-controller-manager-687df44cdb-26dhq\" (UID: \"0d04d185-02c6-4688-83cd-e9f1744776f1\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.846350 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.853944 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.855256 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.859045 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860473 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c54k\" (UniqueName: \"kubernetes.io/projected/88a9a65b-d7fe-4471-b408-562f5c98bdbb-kube-api-access-7c54k\") pod \"mariadb-operator-controller-manager-5777b4f897-8ttcw\" (UID: \"88a9a65b-d7fe-4471-b408-562f5c98bdbb\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860533 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxdcn\" (UniqueName: \"kubernetes.io/projected/fcb8eabd-193f-4b03-b011-5e18514cc858-kube-api-access-xxdcn\") pod \"neutron-operator-controller-manager-797d478b46-kz58p\" (UID: \"fcb8eabd-193f-4b03-b011-5e18514cc858\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860596 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgdzq\" (UniqueName: \"kubernetes.io/projected/6c78484d-5fa5-4f18-ace9-a6811363dca7-kube-api-access-zgdzq\") pod \"ovn-operator-controller-manager-6f96f8c84-xphlj\" (UID: \"6c78484d-5fa5-4f18-ace9-a6811363dca7\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860664 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n2j8\" (UniqueName: \"kubernetes.io/projected/18e70210-0a78-4ea0-8d8b-a955cdbec6b8-kube-api-access-6n2j8\") pod \"manila-operator-controller-manager-59578bc799-2knwl\" (UID: \"18e70210-0a78-4ea0-8d8b-a955cdbec6b8\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860713 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5wx8\" (UniqueName: \"kubernetes.io/projected/c4fa74e8-0166-4704-8a06-401b53e1cefe-kube-api-access-p5wx8\") pod \"octavia-operator-controller-manager-6d7c7ddf95-ffks5\" (UID: \"c4fa74e8-0166-4704-8a06-401b53e1cefe\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.860739 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnl4x\" (UniqueName: \"kubernetes.io/projected/ae4070fd-f49f-403d-a0b5-3285f235a221-kube-api-access-mnl4x\") pod \"nova-operator-controller-manager-57bb74c7bf-v67sz\" (UID: \"ae4070fd-f49f-403d-a0b5-3285f235a221\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.863750 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-vlrbk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.863983 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.882043 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.884164 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnl4x\" (UniqueName: \"kubernetes.io/projected/ae4070fd-f49f-403d-a0b5-3285f235a221-kube-api-access-mnl4x\") pod \"nova-operator-controller-manager-57bb74c7bf-v67sz\" (UID: \"ae4070fd-f49f-403d-a0b5-3285f235a221\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.905221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c54k\" (UniqueName: \"kubernetes.io/projected/88a9a65b-d7fe-4471-b408-562f5c98bdbb-kube-api-access-7c54k\") pod \"mariadb-operator-controller-manager-5777b4f897-8ttcw\" (UID: \"88a9a65b-d7fe-4471-b408-562f5c98bdbb\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.911614 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.912707 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n2j8\" (UniqueName: \"kubernetes.io/projected/18e70210-0a78-4ea0-8d8b-a955cdbec6b8-kube-api-access-6n2j8\") pod \"manila-operator-controller-manager-59578bc799-2knwl\" (UID: \"18e70210-0a78-4ea0-8d8b-a955cdbec6b8\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.919585 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.921101 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxdcn\" (UniqueName: \"kubernetes.io/projected/fcb8eabd-193f-4b03-b011-5e18514cc858-kube-api-access-xxdcn\") pod \"neutron-operator-controller-manager-797d478b46-kz58p\" (UID: \"fcb8eabd-193f-4b03-b011-5e18514cc858\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.921439 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.927708 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.939626 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-bpq2q" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.953069 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.954911 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.958399 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-79qbt" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.959311 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.961485 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.962772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgdzq\" (UniqueName: \"kubernetes.io/projected/6c78484d-5fa5-4f18-ace9-a6811363dca7-kube-api-access-zgdzq\") pod \"ovn-operator-controller-manager-6f96f8c84-xphlj\" (UID: \"6c78484d-5fa5-4f18-ace9-a6811363dca7\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.962885 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.963486 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vq7d\" (UniqueName: \"kubernetes.io/projected/4099adc7-9a49-4191-bfc3-3f17b15312db-kube-api-access-6vq7d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.963519 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5wx8\" (UniqueName: \"kubernetes.io/projected/c4fa74e8-0166-4704-8a06-401b53e1cefe-kube-api-access-p5wx8\") pod \"octavia-operator-controller-manager-6d7c7ddf95-ffks5\" (UID: \"c4fa74e8-0166-4704-8a06-401b53e1cefe\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.964656 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-xjd2r" Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.975749 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc"] Oct 09 08:32:44 crc kubenswrapper[4872]: I1009 08:32:44.997045 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.009376 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgdzq\" (UniqueName: \"kubernetes.io/projected/6c78484d-5fa5-4f18-ace9-a6811363dca7-kube-api-access-zgdzq\") pod \"ovn-operator-controller-manager-6f96f8c84-xphlj\" (UID: \"6c78484d-5fa5-4f18-ace9-a6811363dca7\") " pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.012683 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.020669 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.021042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5wx8\" (UniqueName: \"kubernetes.io/projected/c4fa74e8-0166-4704-8a06-401b53e1cefe-kube-api-access-p5wx8\") pod \"octavia-operator-controller-manager-6d7c7ddf95-ffks5\" (UID: \"c4fa74e8-0166-4704-8a06-401b53e1cefe\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.036357 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.036406 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.036506 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.036582 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.041030 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-572fs" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.055626 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.055862 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.057525 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.058355 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.061816 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9drm9" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.064324 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.064541 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mcsj\" (UniqueName: \"kubernetes.io/projected/a9091168-5676-4cd3-9a0d-f76d0c7c88df-kube-api-access-8mcsj\") pod \"telemetry-operator-controller-manager-578874c84d-b9zg2\" (UID: \"a9091168-5676-4cd3-9a0d-f76d0c7c88df\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.064698 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vq7d\" (UniqueName: \"kubernetes.io/projected/4099adc7-9a49-4191-bfc3-3f17b15312db-kube-api-access-6vq7d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.064840 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kw5s\" (UniqueName: \"kubernetes.io/projected/25798330-695b-4a9d-b271-48b6d025a3f8-kube-api-access-4kw5s\") pod \"placement-operator-controller-manager-664664cb68-sbzmh\" (UID: \"25798330-695b-4a9d-b271-48b6d025a3f8\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.064953 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2lkmr\" (UniqueName: \"kubernetes.io/projected/bfaf85e3-55b3-4564-9ffc-3b80e7f32c85-kube-api-access-2lkmr\") pod \"swift-operator-controller-manager-5f4d5dfdc6-w6jlm\" (UID: \"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.065006 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.065176 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert podName:4099adc7-9a49-4191-bfc3-3f17b15312db nodeName:}" failed. No retries permitted until 2025-10-09 08:32:45.56515866 +0000 UTC m=+923.755687286 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" (UID: "4099adc7-9a49-4191-bfc3-3f17b15312db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.099045 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.100600 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.101615 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.102531 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vq7d\" (UniqueName: \"kubernetes.io/projected/4099adc7-9a49-4191-bfc3-3f17b15312db-kube-api-access-6vq7d\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.107262 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.108923 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-cbdtt" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.116281 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.126954 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.153020 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.164036 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.168110 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mcsj\" (UniqueName: \"kubernetes.io/projected/a9091168-5676-4cd3-9a0d-f76d0c7c88df-kube-api-access-8mcsj\") pod \"telemetry-operator-controller-manager-578874c84d-b9zg2\" (UID: \"a9091168-5676-4cd3-9a0d-f76d0c7c88df\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.168193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8q5hf\" (UniqueName: \"kubernetes.io/projected/3b929d55-693e-419d-ad6e-aa0005f6538a-kube-api-access-8q5hf\") pod \"watcher-operator-controller-manager-646675d848-sjx2l\" (UID: \"3b929d55-693e-419d-ad6e-aa0005f6538a\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.168253 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kw5s\" (UniqueName: \"kubernetes.io/projected/25798330-695b-4a9d-b271-48b6d025a3f8-kube-api-access-4kw5s\") pod \"placement-operator-controller-manager-664664cb68-sbzmh\" (UID: \"25798330-695b-4a9d-b271-48b6d025a3f8\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.168282 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2lkmr\" (UniqueName: \"kubernetes.io/projected/bfaf85e3-55b3-4564-9ffc-3b80e7f32c85-kube-api-access-2lkmr\") pod \"swift-operator-controller-manager-5f4d5dfdc6-w6jlm\" (UID: \"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.169252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.169761 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br7c5\" (UniqueName: \"kubernetes.io/projected/4dfe7dfe-548f-4ec8-a435-72217082dd3b-kube-api-access-br7c5\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.169789 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ck5x\" (UniqueName: \"kubernetes.io/projected/0cf7b66d-2680-43e9-812b-1a3713eea708-kube-api-access-9ck5x\") pod \"test-operator-controller-manager-ffcdd6c94-ppxzv\" (UID: \"0cf7b66d-2680-43e9-812b-1a3713eea708\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.174890 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.177744 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.182611 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-fxvt4" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.188787 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.200104 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.217088 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mcsj\" (UniqueName: \"kubernetes.io/projected/a9091168-5676-4cd3-9a0d-f76d0c7c88df-kube-api-access-8mcsj\") pod \"telemetry-operator-controller-manager-578874c84d-b9zg2\" (UID: \"a9091168-5676-4cd3-9a0d-f76d0c7c88df\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.217169 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2lkmr\" (UniqueName: \"kubernetes.io/projected/bfaf85e3-55b3-4564-9ffc-3b80e7f32c85-kube-api-access-2lkmr\") pod \"swift-operator-controller-manager-5f4d5dfdc6-w6jlm\" (UID: \"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.221820 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kw5s\" (UniqueName: \"kubernetes.io/projected/25798330-695b-4a9d-b271-48b6d025a3f8-kube-api-access-4kw5s\") pod \"placement-operator-controller-manager-664664cb68-sbzmh\" (UID: \"25798330-695b-4a9d-b271-48b6d025a3f8\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.253565 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272293 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ck5x\" (UniqueName: \"kubernetes.io/projected/0cf7b66d-2680-43e9-812b-1a3713eea708-kube-api-access-9ck5x\") pod \"test-operator-controller-manager-ffcdd6c94-ppxzv\" (UID: \"0cf7b66d-2680-43e9-812b-1a3713eea708\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272370 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br7c5\" (UniqueName: \"kubernetes.io/projected/4dfe7dfe-548f-4ec8-a435-72217082dd3b-kube-api-access-br7c5\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272428 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272446 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8q5hf\" (UniqueName: \"kubernetes.io/projected/3b929d55-693e-419d-ad6e-aa0005f6538a-kube-api-access-8q5hf\") pod \"watcher-operator-controller-manager-646675d848-sjx2l\" (UID: \"3b929d55-693e-419d-ad6e-aa0005f6538a\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.272477 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2r7s\" (UniqueName: \"kubernetes.io/projected/add3400b-1fe2-4b5e-a810-7cca5a38ac96-kube-api-access-x2r7s\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz\" (UID: \"add3400b-1fe2-4b5e-a810-7cca5a38ac96\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.272496 4872 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.272567 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert podName:4dfe7dfe-548f-4ec8-a435-72217082dd3b nodeName:}" failed. No retries permitted until 2025-10-09 08:32:45.772546787 +0000 UTC m=+923.963075413 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert") pod "openstack-operator-controller-manager-5f665bd66d-9lhnp" (UID: "4dfe7dfe-548f-4ec8-a435-72217082dd3b") : secret "webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.272923 4872 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.272975 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert podName:57a99d6c-df10-4de9-9434-471ceb56878d nodeName:}" failed. No retries permitted until 2025-10-09 08:32:46.272956369 +0000 UTC m=+924.463484995 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert") pod "infra-operator-controller-manager-585fc5b659-kgzkf" (UID: "57a99d6c-df10-4de9-9434-471ceb56878d") : secret "infra-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.308012 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8q5hf\" (UniqueName: \"kubernetes.io/projected/3b929d55-693e-419d-ad6e-aa0005f6538a-kube-api-access-8q5hf\") pod \"watcher-operator-controller-manager-646675d848-sjx2l\" (UID: \"3b929d55-693e-419d-ad6e-aa0005f6538a\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.316984 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br7c5\" (UniqueName: \"kubernetes.io/projected/4dfe7dfe-548f-4ec8-a435-72217082dd3b-kube-api-access-br7c5\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.320846 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ck5x\" (UniqueName: \"kubernetes.io/projected/0cf7b66d-2680-43e9-812b-1a3713eea708-kube-api-access-9ck5x\") pod \"test-operator-controller-manager-ffcdd6c94-ppxzv\" (UID: \"0cf7b66d-2680-43e9-812b-1a3713eea708\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.321128 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:45 crc kubenswrapper[4872]: W1009 08:32:45.371388 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod283f56cf_99e9_471b_8f40_10ec0f31ad06.slice/crio-e45a95dcdb83b67ec166a958bd712e3bf396f66302b9b4450bb3fe447ec09d07 WatchSource:0}: Error finding container e45a95dcdb83b67ec166a958bd712e3bf396f66302b9b4450bb3fe447ec09d07: Status 404 returned error can't find the container with id e45a95dcdb83b67ec166a958bd712e3bf396f66302b9b4450bb3fe447ec09d07 Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.373738 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2r7s\" (UniqueName: \"kubernetes.io/projected/add3400b-1fe2-4b5e-a810-7cca5a38ac96-kube-api-access-x2r7s\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz\" (UID: \"add3400b-1fe2-4b5e-a810-7cca5a38ac96\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.409241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2r7s\" (UniqueName: \"kubernetes.io/projected/add3400b-1fe2-4b5e-a810-7cca5a38ac96-kube-api-access-x2r7s\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz\" (UID: \"add3400b-1fe2-4b5e-a810-7cca5a38ac96\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.416158 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.451129 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.469787 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.492243 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.510114 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.579259 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.579558 4872 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: E1009 08:32:45.579612 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert podName:4099adc7-9a49-4191-bfc3-3f17b15312db nodeName:}" failed. No retries permitted until 2025-10-09 08:32:46.579595021 +0000 UTC m=+924.770123647 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" (UID: "4099adc7-9a49-4191-bfc3-3f17b15312db") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.665017 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.782300 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.788331 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4dfe7dfe-548f-4ec8-a435-72217082dd3b-cert\") pod \"openstack-operator-controller-manager-5f665bd66d-9lhnp\" (UID: \"4dfe7dfe-548f-4ec8-a435-72217082dd3b\") " pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.798298 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.839984 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.871134 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.872978 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" event={"ID":"91db56d0-f7bc-4d02-92ec-c7ead6625176","Type":"ContainerStarted","Data":"e0127eee3c1b4b0e6ccc964073c8f9b5a6be7fd85497f85f56a89cbc717ba977"} Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.875820 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" event={"ID":"283f56cf-99e9-471b-8f40-10ec0f31ad06","Type":"ContainerStarted","Data":"e45a95dcdb83b67ec166a958bd712e3bf396f66302b9b4450bb3fe447ec09d07"} Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.881907 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.894865 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4"] Oct 09 08:32:45 crc kubenswrapper[4872]: I1009 08:32:45.897761 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:45 crc kubenswrapper[4872]: W1009 08:32:45.956869 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9b5d69d_fb26_4286_a9b0_3b00d042972a.slice/crio-86ed7892598baf462a044f4ce8f9c6fa5de33fe09847bd0240ce5423e9466ac2 WatchSource:0}: Error finding container 86ed7892598baf462a044f4ce8f9c6fa5de33fe09847bd0240ce5423e9466ac2: Status 404 returned error can't find the container with id 86ed7892598baf462a044f4ce8f9c6fa5de33fe09847bd0240ce5423e9466ac2 Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.025116 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.037268 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.050435 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p"] Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.058715 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5d3f64a_0214_457c_b4c6_f48beabe8669.slice/crio-a1ca1f86608c4c10f4965c04882fa21d17b645e59749da1069bcb1e4383026f9 WatchSource:0}: Error finding container a1ca1f86608c4c10f4965c04882fa21d17b645e59749da1069bcb1e4383026f9: Status 404 returned error can't find the container with id a1ca1f86608c4c10f4965c04882fa21d17b645e59749da1069bcb1e4383026f9 Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.074799 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.217248 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-2knwl"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.225681 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.244789 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj"] Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.258336 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c78484d_5fa5_4f18_ace9_a6811363dca7.slice/crio-ce2bce6286ee44a0b224f0b4ef6e9f0552d4c3ab256f41a162e8226b9a454069 WatchSource:0}: Error finding container ce2bce6286ee44a0b224f0b4ef6e9f0552d4c3ab256f41a162e8226b9a454069: Status 404 returned error can't find the container with id ce2bce6286ee44a0b224f0b4ef6e9f0552d4c3ab256f41a162e8226b9a454069 Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.293358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.297388 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/57a99d6c-df10-4de9-9434-471ceb56878d-cert\") pod \"infra-operator-controller-manager-585fc5b659-kgzkf\" (UID: \"57a99d6c-df10-4de9-9434-471ceb56878d\") " pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.377630 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.385374 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv"] Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.386958 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9091168_5676_4cd3_9a0d_f76d0c7c88df.slice/crio-5ba60f715def3a0c7c90fbc3824b24ffb55079e91c98f0daa5b827e19663d0fa WatchSource:0}: Error finding container 5ba60f715def3a0c7c90fbc3824b24ffb55079e91c98f0daa5b827e19663d0fa: Status 404 returned error can't find the container with id 5ba60f715def3a0c7c90fbc3824b24ffb55079e91c98f0daa5b827e19663d0fa Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.388939 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b929d55_693e_419d_ad6e_aa0005f6538a.slice/crio-e6f04abac25619e7f520e047b465985a820505ebb651de052b5cb6f689df6aff WatchSource:0}: Error finding container e6f04abac25619e7f520e047b465985a820505ebb651de052b5cb6f689df6aff: Status 404 returned error can't find the container with id e6f04abac25619e7f520e047b465985a820505ebb651de052b5cb6f689df6aff Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.390016 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8mcsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-578874c84d-b9zg2_openstack-operators(a9091168-5676-4cd3-9a0d-f76d0c7c88df): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.394861 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cf7b66d_2680_43e9_812b_1a3713eea708.slice/crio-699b8e77fd73aa2f2cf8fa9fac136d1fde509731354902fb4b7d749e5efdfb5b WatchSource:0}: Error finding container 699b8e77fd73aa2f2cf8fa9fac136d1fde509731354902fb4b7d749e5efdfb5b: Status 404 returned error can't find the container with id 699b8e77fd73aa2f2cf8fa9fac136d1fde509731354902fb4b7d749e5efdfb5b Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.403685 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9ck5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-ppxzv_openstack-operators(0cf7b66d-2680-43e9-812b-1a3713eea708): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.402781 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8q5hf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-sjx2l_openstack-operators(3b929d55-693e-419d-ad6e-aa0005f6538a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.404769 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.411522 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh"] Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.415087 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4kw5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-sbzmh_openstack-operators(25798330-695b-4a9d-b271-48b6d025a3f8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.453259 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.472066 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm"] Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.480259 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz"] Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.496687 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfaf85e3_55b3_4564_9ffc_3b80e7f32c85.slice/crio-6e8dfdff76c126bab333c5c0ad34960818563d92e0c367a20f3d445b965f95f4 WatchSource:0}: Error finding container 6e8dfdff76c126bab333c5c0ad34960818563d92e0c367a20f3d445b965f95f4: Status 404 returned error can't find the container with id 6e8dfdff76c126bab333c5c0ad34960818563d92e0c367a20f3d445b965f95f4 Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.504799 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x2r7s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz_openstack-operators(add3400b-1fe2-4b5e-a810-7cca5a38ac96): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.506505 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" podUID="add3400b-1fe2-4b5e-a810-7cca5a38ac96" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.525870 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp"] Oct 09 08:32:46 crc kubenswrapper[4872]: W1009 08:32:46.537230 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dfe7dfe_548f_4ec8_a435_72217082dd3b.slice/crio-d43121c560278c672be73dc1aed040cda0a736a1afb61cfd302d47388a26988c WatchSource:0}: Error finding container d43121c560278c672be73dc1aed040cda0a736a1afb61cfd302d47388a26988c: Status 404 returned error can't find the container with id d43121c560278c672be73dc1aed040cda0a736a1afb61cfd302d47388a26988c Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.570086 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" podUID="a9091168-5676-4cd3-9a0d-f76d0c7c88df" Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.575472 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" podUID="3b929d55-693e-419d-ad6e-aa0005f6538a" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.601110 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.612343 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4099adc7-9a49-4191-bfc3-3f17b15312db-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc\" (UID: \"4099adc7-9a49-4191-bfc3-3f17b15312db\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.668312 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" podUID="0cf7b66d-2680-43e9-812b-1a3713eea708" Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.770665 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" podUID="25798330-695b-4a9d-b271-48b6d025a3f8" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.883742 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.889749 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" event={"ID":"0cf7b66d-2680-43e9-812b-1a3713eea708","Type":"ContainerStarted","Data":"109a03755f2eadd13a0acfb764131be47b7f2ce8080e22628d073fde504564fd"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.889806 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" event={"ID":"0cf7b66d-2680-43e9-812b-1a3713eea708","Type":"ContainerStarted","Data":"699b8e77fd73aa2f2cf8fa9fac136d1fde509731354902fb4b7d749e5efdfb5b"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.894716 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" event={"ID":"deb85a9b-515a-4027-a178-c83829fd5a34","Type":"ContainerStarted","Data":"f8749e575678c3d0696cb42f8e3a392eab738c2e89c948056996eae6c2674bb4"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.896074 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" event={"ID":"18e70210-0a78-4ea0-8d8b-a955cdbec6b8","Type":"ContainerStarted","Data":"a69a438b1480cece0850e53550154a4a253efb03a681d54c34b6ac79625eed46"} Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.896940 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" podUID="0cf7b66d-2680-43e9-812b-1a3713eea708" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.899332 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" event={"ID":"4dfe7dfe-548f-4ec8-a435-72217082dd3b","Type":"ContainerStarted","Data":"e435ffc5a5753c9aac92908608862d9a469942b67ab490b8e9cd0f978b4140e1"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.899373 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" event={"ID":"4dfe7dfe-548f-4ec8-a435-72217082dd3b","Type":"ContainerStarted","Data":"d43121c560278c672be73dc1aed040cda0a736a1afb61cfd302d47388a26988c"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.900738 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" event={"ID":"25798330-695b-4a9d-b271-48b6d025a3f8","Type":"ContainerStarted","Data":"bfb5987d9bc16eb8ed21a5c52e512c8aaedf92baedbea0ad238a8bd4c64092cc"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.900767 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" event={"ID":"25798330-695b-4a9d-b271-48b6d025a3f8","Type":"ContainerStarted","Data":"5d6ceb261620028e7413ef31f6990928f01b7957ab24ec96bb600b57f3003a32"} Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.902543 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" podUID="25798330-695b-4a9d-b271-48b6d025a3f8" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.903624 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" event={"ID":"fab5d9a4-352a-4092-9139-c74242142eda","Type":"ContainerStarted","Data":"095b6b8b27c7ed915db2b9533a14064422274afbb8e721282a67c71e24b7712d"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.917199 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" event={"ID":"a34a57a1-2b91-4d65-8336-4a363b33dd45","Type":"ContainerStarted","Data":"a6ed4becd10869fcd91fe4705155bcfa71f95fdff492bc3f2dee6f8b7f886349"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.928460 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" event={"ID":"b9b5d69d-fb26-4286-a9b0-3b00d042972a","Type":"ContainerStarted","Data":"86ed7892598baf462a044f4ce8f9c6fa5de33fe09847bd0240ce5423e9466ac2"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.933397 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" event={"ID":"add3400b-1fe2-4b5e-a810-7cca5a38ac96","Type":"ContainerStarted","Data":"e0a7a188d0a5c1166ae82a8b4d8cae9cefc6a936e2b7932010a2640f7d93bd4f"} Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.940670 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" podUID="add3400b-1fe2-4b5e-a810-7cca5a38ac96" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.947137 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" event={"ID":"a9091168-5676-4cd3-9a0d-f76d0c7c88df","Type":"ContainerStarted","Data":"ef1d0fcc92444eafec722283a138227a40a086fd3a43040b596c0c21e2f43443"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.947266 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" event={"ID":"a9091168-5676-4cd3-9a0d-f76d0c7c88df","Type":"ContainerStarted","Data":"5ba60f715def3a0c7c90fbc3824b24ffb55079e91c98f0daa5b827e19663d0fa"} Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.948592 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" podUID="a9091168-5676-4cd3-9a0d-f76d0c7c88df" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.950038 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" event={"ID":"3b929d55-693e-419d-ad6e-aa0005f6538a","Type":"ContainerStarted","Data":"602e086b21a341e01b49dcf38c928bf58ac362dd4398a57b47235a8a0b0bd710"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.950074 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" event={"ID":"3b929d55-693e-419d-ad6e-aa0005f6538a","Type":"ContainerStarted","Data":"e6f04abac25619e7f520e047b465985a820505ebb651de052b5cb6f689df6aff"} Oct 09 08:32:46 crc kubenswrapper[4872]: E1009 08:32:46.951121 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" podUID="3b929d55-693e-419d-ad6e-aa0005f6538a" Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.955064 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" event={"ID":"6c78484d-5fa5-4f18-ace9-a6811363dca7","Type":"ContainerStarted","Data":"ce2bce6286ee44a0b224f0b4ef6e9f0552d4c3ab256f41a162e8226b9a454069"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.963988 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" event={"ID":"fcb8eabd-193f-4b03-b011-5e18514cc858","Type":"ContainerStarted","Data":"506bd337faf5aa90c5eb9c2fe73667e7476348ae45dad3529922cbcbcce63ad9"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.974394 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" event={"ID":"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85","Type":"ContainerStarted","Data":"6e8dfdff76c126bab333c5c0ad34960818563d92e0c367a20f3d445b965f95f4"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.980352 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" event={"ID":"c4fa74e8-0166-4704-8a06-401b53e1cefe","Type":"ContainerStarted","Data":"a52d0fd58b88deb6991ba852bd8396584cdf541db8a0b550a14bc3027b798bcd"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.982263 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" event={"ID":"ae4070fd-f49f-403d-a0b5-3285f235a221","Type":"ContainerStarted","Data":"2469421e40f84afb19393662108959f317c4c271b1df354c8e4829a5b56c69d0"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.984529 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" event={"ID":"88a9a65b-d7fe-4471-b408-562f5c98bdbb","Type":"ContainerStarted","Data":"49d24c5ce6d5a30ed8299ce468553c1a73df99fbd100fe00f6cc2d74dfe7975b"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.989136 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" event={"ID":"c5d3f64a-0214-457c-b4c6-f48beabe8669","Type":"ContainerStarted","Data":"a1ca1f86608c4c10f4965c04882fa21d17b645e59749da1069bcb1e4383026f9"} Oct 09 08:32:46 crc kubenswrapper[4872]: I1009 08:32:46.998076 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" event={"ID":"0d04d185-02c6-4688-83cd-e9f1744776f1","Type":"ContainerStarted","Data":"705339eee03ec2cfcac3dc4b6bd15a17ba6bded45be3ff620b5307730e81adbe"} Oct 09 08:32:47 crc kubenswrapper[4872]: I1009 08:32:47.029257 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf"] Oct 09 08:32:47 crc kubenswrapper[4872]: W1009 08:32:47.062047 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57a99d6c_df10_4de9_9434_471ceb56878d.slice/crio-405c539bfe2c0cd007dc4b7f4957bf028e89d9d457be88224e5383ba66316015 WatchSource:0}: Error finding container 405c539bfe2c0cd007dc4b7f4957bf028e89d9d457be88224e5383ba66316015: Status 404 returned error can't find the container with id 405c539bfe2c0cd007dc4b7f4957bf028e89d9d457be88224e5383ba66316015 Oct 09 08:32:47 crc kubenswrapper[4872]: I1009 08:32:47.448808 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc"] Oct 09 08:32:47 crc kubenswrapper[4872]: W1009 08:32:47.461447 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4099adc7_9a49_4191_bfc3_3f17b15312db.slice/crio-75a0217075bef19f908fe75e8fea7ba0177adb8debd2a2ef01cf552601a5b88f WatchSource:0}: Error finding container 75a0217075bef19f908fe75e8fea7ba0177adb8debd2a2ef01cf552601a5b88f: Status 404 returned error can't find the container with id 75a0217075bef19f908fe75e8fea7ba0177adb8debd2a2ef01cf552601a5b88f Oct 09 08:32:48 crc kubenswrapper[4872]: I1009 08:32:48.010383 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" event={"ID":"57a99d6c-df10-4de9-9434-471ceb56878d","Type":"ContainerStarted","Data":"405c539bfe2c0cd007dc4b7f4957bf028e89d9d457be88224e5383ba66316015"} Oct 09 08:32:48 crc kubenswrapper[4872]: I1009 08:32:48.014139 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" event={"ID":"4dfe7dfe-548f-4ec8-a435-72217082dd3b","Type":"ContainerStarted","Data":"61201c63678319d3a330230d243a282adba9a830e97b63640e487119ae1a6ff0"} Oct 09 08:32:48 crc kubenswrapper[4872]: I1009 08:32:48.014474 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:48 crc kubenswrapper[4872]: I1009 08:32:48.017276 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" event={"ID":"4099adc7-9a49-4191-bfc3-3f17b15312db","Type":"ContainerStarted","Data":"75a0217075bef19f908fe75e8fea7ba0177adb8debd2a2ef01cf552601a5b88f"} Oct 09 08:32:48 crc kubenswrapper[4872]: E1009 08:32:48.022945 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" podUID="3b929d55-693e-419d-ad6e-aa0005f6538a" Oct 09 08:32:48 crc kubenswrapper[4872]: E1009 08:32:48.022978 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" podUID="add3400b-1fe2-4b5e-a810-7cca5a38ac96" Oct 09 08:32:48 crc kubenswrapper[4872]: E1009 08:32:48.022968 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" podUID="a9091168-5676-4cd3-9a0d-f76d0c7c88df" Oct 09 08:32:48 crc kubenswrapper[4872]: E1009 08:32:48.023043 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" podUID="0cf7b66d-2680-43e9-812b-1a3713eea708" Oct 09 08:32:48 crc kubenswrapper[4872]: E1009 08:32:48.023055 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" podUID="25798330-695b-4a9d-b271-48b6d025a3f8" Oct 09 08:32:48 crc kubenswrapper[4872]: I1009 08:32:48.039660 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" podStartSLOduration=3.039626429 podStartE2EDuration="3.039626429s" podCreationTimestamp="2025-10-09 08:32:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:32:48.036046606 +0000 UTC m=+926.226575252" watchObservedRunningTime="2025-10-09 08:32:48.039626429 +0000 UTC m=+926.230155055" Oct 09 08:32:55 crc kubenswrapper[4872]: I1009 08:32:55.905696 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5f665bd66d-9lhnp" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.092989 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" event={"ID":"c5d3f64a-0214-457c-b4c6-f48beabe8669","Type":"ContainerStarted","Data":"d33676efe907228f99a608535ed9169d4dd6a961e7568318f66b2a1c8561ba74"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.094423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" event={"ID":"b9b5d69d-fb26-4286-a9b0-3b00d042972a","Type":"ContainerStarted","Data":"e4ca07904a023cdc2aedd696c9636f9ad4ce6be9bd6a37c1952aadfdd522000f"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.100508 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" event={"ID":"18e70210-0a78-4ea0-8d8b-a955cdbec6b8","Type":"ContainerStarted","Data":"52373ef661a93b28740451161bccb72a8c766aaf07e6c68c221b220a7f30d9cd"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.102193 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" event={"ID":"0d04d185-02c6-4688-83cd-e9f1744776f1","Type":"ContainerStarted","Data":"38b3114ef35ed5d448a635e704d57dc4c48c894b60e93a74367a1a9a4c10bca1"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.106063 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" event={"ID":"ae4070fd-f49f-403d-a0b5-3285f235a221","Type":"ContainerStarted","Data":"7c5b4d0703ee28bcd4bcd73cb66026de4f7063d3b612b28a64cc980ae5409765"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.106115 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" event={"ID":"ae4070fd-f49f-403d-a0b5-3285f235a221","Type":"ContainerStarted","Data":"bbca1be37cf85bdebb55c6bdf98e3b2a10f894e72eed824fb19d5d685780951f"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.107149 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.109077 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" event={"ID":"deb85a9b-515a-4027-a178-c83829fd5a34","Type":"ContainerStarted","Data":"e9ef0c549a903d276b50f00964d7262c5538986bf4e83ae301863025ab54bd3b"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.141056 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" event={"ID":"57a99d6c-df10-4de9-9434-471ceb56878d","Type":"ContainerStarted","Data":"ee0bb944e40139f63ccf8a6e77ae93959c3de4a064ed14a68bead03884de144c"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.164866 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" event={"ID":"283f56cf-99e9-471b-8f40-10ec0f31ad06","Type":"ContainerStarted","Data":"c5f3b789e4de579a489e0c7dd324df339210728edc880996b2ee53b0401b1d95"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.174964 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" event={"ID":"6c78484d-5fa5-4f18-ace9-a6811363dca7","Type":"ContainerStarted","Data":"e1af820c68091836a28b93468f4aee287719194c5773998ca5f28802afbf4f64"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.219125 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" event={"ID":"fcb8eabd-193f-4b03-b011-5e18514cc858","Type":"ContainerStarted","Data":"1ae180d4028e5004a9520d68263fc56b0b552c917c168dd28dedb049c22c7ebe"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.237278 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" event={"ID":"91db56d0-f7bc-4d02-92ec-c7ead6625176","Type":"ContainerStarted","Data":"c3d1d9a14a0d2b8955ad0029eda1a7a1f37bba1b59865f7ddd79ae2354cc7093"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.237339 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" event={"ID":"91db56d0-f7bc-4d02-92ec-c7ead6625176","Type":"ContainerStarted","Data":"dec28b2cef2b6bcce1ae0429f651ddf63e270c70014d342849d6f49cad2d787d"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.238728 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.247871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" event={"ID":"fab5d9a4-352a-4092-9139-c74242142eda","Type":"ContainerStarted","Data":"e7fc9a539f998281893226ad817ef77f29616b3405cc446cec4610519c542627"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.247919 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" event={"ID":"fab5d9a4-352a-4092-9139-c74242142eda","Type":"ContainerStarted","Data":"b40554dc2061edbeb733f672344303c7a67d2a1fa8506deb1f2359562f80136e"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.248524 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.258049 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" event={"ID":"88a9a65b-d7fe-4471-b408-562f5c98bdbb","Type":"ContainerStarted","Data":"bb5291000ce90713e0ac67f21a6a37d1f5fabe2d8f75e27803caa294725a98b3"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.263166 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" podStartSLOduration=2.8739866530000002 podStartE2EDuration="13.263149463s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.208335389 +0000 UTC m=+923.398864015" lastFinishedPulling="2025-10-09 08:32:55.597498199 +0000 UTC m=+933.788026825" observedRunningTime="2025-10-09 08:32:57.262680119 +0000 UTC m=+935.453208765" watchObservedRunningTime="2025-10-09 08:32:57.263149463 +0000 UTC m=+935.453678089" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.270317 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" podStartSLOduration=3.503092594 podStartE2EDuration="13.270302739s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.88922766 +0000 UTC m=+924.079756286" lastFinishedPulling="2025-10-09 08:32:55.656437805 +0000 UTC m=+933.846966431" observedRunningTime="2025-10-09 08:32:57.176407537 +0000 UTC m=+935.366936163" watchObservedRunningTime="2025-10-09 08:32:57.270302739 +0000 UTC m=+935.460831365" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.271844 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" event={"ID":"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85","Type":"ContainerStarted","Data":"f3fa2e52dce5245c6551f46dfca33189b0b69e71f544f5d2b7ed3f26a6cb54d5"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.278701 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" event={"ID":"a34a57a1-2b91-4d65-8336-4a363b33dd45","Type":"ContainerStarted","Data":"e10fda72acc0fbd1320a87b4a2cc7ccd3d24d2385cd3239dd56ea3579834da3e"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.298923 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" podStartSLOduration=3.5034817350000003 podStartE2EDuration="13.298905532s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.890318951 +0000 UTC m=+924.080847577" lastFinishedPulling="2025-10-09 08:32:55.685742748 +0000 UTC m=+933.876271374" observedRunningTime="2025-10-09 08:32:57.296319337 +0000 UTC m=+935.486847963" watchObservedRunningTime="2025-10-09 08:32:57.298905532 +0000 UTC m=+935.489434168" Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.301314 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" event={"ID":"c4fa74e8-0166-4704-8a06-401b53e1cefe","Type":"ContainerStarted","Data":"ac1e7db2d7d7fc812281211836c64ce895174165dc71ae2de4236a89352cbefb"} Oct 09 08:32:57 crc kubenswrapper[4872]: I1009 08:32:57.325852 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" event={"ID":"4099adc7-9a49-4191-bfc3-3f17b15312db","Type":"ContainerStarted","Data":"96afd72e7c316ddd97b2b78050f0dcca439dedf554caf05ce644761a3e6961fe"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.340012 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" event={"ID":"deb85a9b-515a-4027-a178-c83829fd5a34","Type":"ContainerStarted","Data":"41b0cf48c9f55d0649dbad5721b5aa721aa22b40f947a2116ceb4cd251c5eb68"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.340259 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.344533 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" event={"ID":"18e70210-0a78-4ea0-8d8b-a955cdbec6b8","Type":"ContainerStarted","Data":"f780e377df81c811bc30fffae177edb8b7db8e076f8394ca03baab068195f058"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.344608 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.347233 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" event={"ID":"b9b5d69d-fb26-4286-a9b0-3b00d042972a","Type":"ContainerStarted","Data":"cbb8e3b3098a0f9ac9e20e1bef4e7cf869471e0bb28ac3ac7c312cee28671bfc"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.347364 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.349506 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" event={"ID":"4099adc7-9a49-4191-bfc3-3f17b15312db","Type":"ContainerStarted","Data":"7433bf8d7dbb7326165c75bce4e60f8448c3aee31a8dcf754174d32814e07e75"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.349557 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.352059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" event={"ID":"283f56cf-99e9-471b-8f40-10ec0f31ad06","Type":"ContainerStarted","Data":"1dfbf060c3558e3032b776f91b46d52dafbd727e565d73db0098e11aad0514d2"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.352259 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.354776 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" event={"ID":"88a9a65b-d7fe-4471-b408-562f5c98bdbb","Type":"ContainerStarted","Data":"0fdf450abb0602925042e76ae5e38e6303e392ca3603d2d7679f5b7e0508a8aa"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.354899 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.356586 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" event={"ID":"bfaf85e3-55b3-4564-9ffc-3b80e7f32c85","Type":"ContainerStarted","Data":"171799b25914cc3884ce75e199838c8c491cea38128a210bc925a83e0e676621"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.359455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" event={"ID":"c5d3f64a-0214-457c-b4c6-f48beabe8669","Type":"ContainerStarted","Data":"5616207b06118bbd51fa21131f5b83694ee796c50a115cf1d954c2cce2be159e"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.360022 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.361769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" event={"ID":"0d04d185-02c6-4688-83cd-e9f1744776f1","Type":"ContainerStarted","Data":"9f2b269efa63f8d920bbce91b2db0ea320e671d59579042d590a5088c1051701"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.362242 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.363399 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" podStartSLOduration=4.601523747 podStartE2EDuration="14.363380939s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.886193672 +0000 UTC m=+924.076722298" lastFinishedPulling="2025-10-09 08:32:55.648050864 +0000 UTC m=+933.838579490" observedRunningTime="2025-10-09 08:32:58.359068845 +0000 UTC m=+936.549597481" watchObservedRunningTime="2025-10-09 08:32:58.363380939 +0000 UTC m=+936.553909575" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.364490 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" event={"ID":"6c78484d-5fa5-4f18-ace9-a6811363dca7","Type":"ContainerStarted","Data":"63016cd45cdc694a3f36780fe4a9eb2306408bdff514c81aecba87c37847c25c"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.364754 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.366401 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" event={"ID":"a34a57a1-2b91-4d65-8336-4a363b33dd45","Type":"ContainerStarted","Data":"3dadd2eafc59362fae5d4cdd863e73bc61d9ac1052590e87898d9977831734f6"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.366502 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.367951 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" event={"ID":"fcb8eabd-193f-4b03-b011-5e18514cc858","Type":"ContainerStarted","Data":"c5029c4cc6d13b680a0ba1ca08d98ce657347b53c62dd558249621d7f7295cf7"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.368071 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.369872 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" event={"ID":"57a99d6c-df10-4de9-9434-471ceb56878d","Type":"ContainerStarted","Data":"98fc1fe19820b75e19fe28f457661605ca91014be722be7647026ebc4c8772f4"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.369987 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.372588 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" event={"ID":"c4fa74e8-0166-4704-8a06-401b53e1cefe","Type":"ContainerStarted","Data":"6110d9caf3446eb1a06140d707eff05b73b995345966b38e400cb2fd1437d146"} Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.373489 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.377323 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" podStartSLOduration=4.6993910119999995 podStartE2EDuration="14.377307849s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.96849449 +0000 UTC m=+924.159023116" lastFinishedPulling="2025-10-09 08:32:55.646411327 +0000 UTC m=+933.836939953" observedRunningTime="2025-10-09 08:32:58.372783289 +0000 UTC m=+936.563311925" watchObservedRunningTime="2025-10-09 08:32:58.377307849 +0000 UTC m=+936.567836475" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.400716 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" podStartSLOduration=4.809184572 podStartE2EDuration="14.400689522s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.049279315 +0000 UTC m=+924.239807941" lastFinishedPulling="2025-10-09 08:32:55.640784265 +0000 UTC m=+933.831312891" observedRunningTime="2025-10-09 08:32:58.387363759 +0000 UTC m=+936.577892395" watchObservedRunningTime="2025-10-09 08:32:58.400689522 +0000 UTC m=+936.591218208" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.414228 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" podStartSLOduration=6.165493695 podStartE2EDuration="14.414216701s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:47.464992647 +0000 UTC m=+925.655521273" lastFinishedPulling="2025-10-09 08:32:55.713715653 +0000 UTC m=+933.904244279" observedRunningTime="2025-10-09 08:32:58.409833485 +0000 UTC m=+936.600362141" watchObservedRunningTime="2025-10-09 08:32:58.414216701 +0000 UTC m=+936.604745327" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.432478 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" podStartSLOduration=4.115223806 podStartE2EDuration="14.432462546s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.380845273 +0000 UTC m=+923.571373899" lastFinishedPulling="2025-10-09 08:32:55.698084013 +0000 UTC m=+933.888612639" observedRunningTime="2025-10-09 08:32:58.427821643 +0000 UTC m=+936.618350279" watchObservedRunningTime="2025-10-09 08:32:58.432462546 +0000 UTC m=+936.622991172" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.445067 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" podStartSLOduration=5.001500795 podStartE2EDuration="14.445033728s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.248252489 +0000 UTC m=+924.438781115" lastFinishedPulling="2025-10-09 08:32:55.691785422 +0000 UTC m=+933.882314048" observedRunningTime="2025-10-09 08:32:58.442245388 +0000 UTC m=+936.632774034" watchObservedRunningTime="2025-10-09 08:32:58.445033728 +0000 UTC m=+936.635562364" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.461617 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" podStartSLOduration=5.25469185 podStartE2EDuration="14.461603195s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.504659777 +0000 UTC m=+924.695188403" lastFinishedPulling="2025-10-09 08:32:55.711571122 +0000 UTC m=+933.902099748" observedRunningTime="2025-10-09 08:32:58.455813558 +0000 UTC m=+936.646342184" watchObservedRunningTime="2025-10-09 08:32:58.461603195 +0000 UTC m=+936.652131821" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.480821 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" podStartSLOduration=5.072777406 podStartE2EDuration="14.480790647s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.267124672 +0000 UTC m=+924.457653298" lastFinishedPulling="2025-10-09 08:32:55.675137913 +0000 UTC m=+933.865666539" observedRunningTime="2025-10-09 08:32:58.475077472 +0000 UTC m=+936.665606118" watchObservedRunningTime="2025-10-09 08:32:58.480790647 +0000 UTC m=+936.671319273" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.493609 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" podStartSLOduration=4.688320464 podStartE2EDuration="14.493590125s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:45.864735465 +0000 UTC m=+924.055264091" lastFinishedPulling="2025-10-09 08:32:55.670005126 +0000 UTC m=+933.860533752" observedRunningTime="2025-10-09 08:32:58.489225369 +0000 UTC m=+936.679754005" watchObservedRunningTime="2025-10-09 08:32:58.493590125 +0000 UTC m=+936.684118761" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.526105 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" podStartSLOduration=4.909783425 podStartE2EDuration="14.526081959s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.076276761 +0000 UTC m=+924.266805387" lastFinishedPulling="2025-10-09 08:32:55.692575295 +0000 UTC m=+933.883103921" observedRunningTime="2025-10-09 08:32:58.510636165 +0000 UTC m=+936.701164791" watchObservedRunningTime="2025-10-09 08:32:58.526081959 +0000 UTC m=+936.716610585" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.531976 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" podStartSLOduration=5.93457001 podStartE2EDuration="14.531953608s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:47.071682901 +0000 UTC m=+925.262211537" lastFinishedPulling="2025-10-09 08:32:55.669066509 +0000 UTC m=+933.859595135" observedRunningTime="2025-10-09 08:32:58.529501767 +0000 UTC m=+936.720030393" watchObservedRunningTime="2025-10-09 08:32:58.531953608 +0000 UTC m=+936.722482234" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.545249 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" podStartSLOduration=5.096302532 podStartE2EDuration="14.54523053s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.249226437 +0000 UTC m=+924.439755073" lastFinishedPulling="2025-10-09 08:32:55.698154445 +0000 UTC m=+933.888683071" observedRunningTime="2025-10-09 08:32:58.545026964 +0000 UTC m=+936.735555600" watchObservedRunningTime="2025-10-09 08:32:58.54523053 +0000 UTC m=+936.735759156" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.563014 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" podStartSLOduration=4.943593998 podStartE2EDuration="14.562993931s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.078038642 +0000 UTC m=+924.268567278" lastFinishedPulling="2025-10-09 08:32:55.697438585 +0000 UTC m=+933.887967211" observedRunningTime="2025-10-09 08:32:58.557594805 +0000 UTC m=+936.748123431" watchObservedRunningTime="2025-10-09 08:32:58.562993931 +0000 UTC m=+936.753522557" Oct 09 08:32:58 crc kubenswrapper[4872]: I1009 08:32:58.573648 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" podStartSLOduration=4.967618749 podStartE2EDuration="14.573616916s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.086340911 +0000 UTC m=+924.276869537" lastFinishedPulling="2025-10-09 08:32:55.692339078 +0000 UTC m=+933.882867704" observedRunningTime="2025-10-09 08:32:58.570933689 +0000 UTC m=+936.761462315" watchObservedRunningTime="2025-10-09 08:32:58.573616916 +0000 UTC m=+936.764145542" Oct 09 08:32:59 crc kubenswrapper[4872]: I1009 08:32:59.381133 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:33:00 crc kubenswrapper[4872]: I1009 08:33:00.395192 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" event={"ID":"25798330-695b-4a9d-b271-48b6d025a3f8","Type":"ContainerStarted","Data":"f598bac3e87b328cc2185b7e9915b80dfd8b194d7ca8203489372b019d2d7720"} Oct 09 08:33:00 crc kubenswrapper[4872]: I1009 08:33:00.395671 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:33:00 crc kubenswrapper[4872]: I1009 08:33:00.418392 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" podStartSLOduration=2.942814904 podStartE2EDuration="16.418336672s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.414919905 +0000 UTC m=+924.605448531" lastFinishedPulling="2025-10-09 08:32:59.890441673 +0000 UTC m=+938.080970299" observedRunningTime="2025-10-09 08:33:00.413362189 +0000 UTC m=+938.603890815" watchObservedRunningTime="2025-10-09 08:33:00.418336672 +0000 UTC m=+938.608865298" Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.410560 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" event={"ID":"3b929d55-693e-419d-ad6e-aa0005f6538a","Type":"ContainerStarted","Data":"6a905ad6ff40e31da1d907e25b955cb640b172f49a0089b811b20da292bc7be1"} Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.411102 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.415672 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" event={"ID":"0cf7b66d-2680-43e9-812b-1a3713eea708","Type":"ContainerStarted","Data":"2025f35224c82e278b1c3e322e9a14f814c96e6b5a0f359439b888d14f764618"} Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.415950 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.452673 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" podStartSLOduration=3.262756157 podStartE2EDuration="18.45261401s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.40258289 +0000 UTC m=+924.593111516" lastFinishedPulling="2025-10-09 08:33:01.592440743 +0000 UTC m=+939.782969369" observedRunningTime="2025-10-09 08:33:02.430538545 +0000 UTC m=+940.621067181" watchObservedRunningTime="2025-10-09 08:33:02.45261401 +0000 UTC m=+940.643142646" Oct 09 08:33:02 crc kubenswrapper[4872]: I1009 08:33:02.454385 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" podStartSLOduration=3.257959129 podStartE2EDuration="18.454373871s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.403525997 +0000 UTC m=+924.594054623" lastFinishedPulling="2025-10-09 08:33:01.599940729 +0000 UTC m=+939.790469365" observedRunningTime="2025-10-09 08:33:02.449515171 +0000 UTC m=+940.640043867" watchObservedRunningTime="2025-10-09 08:33:02.454373871 +0000 UTC m=+940.644902537" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.623863 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-lrlzc" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.712101 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-vsdt6" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.829160 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-qnnp4" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.858394 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-n5rbg" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.889580 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-26dhq" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.922559 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-v67sz" Oct 09 08:33:04 crc kubenswrapper[4872]: I1009 08:33:04.924916 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-dskmk" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.039568 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-74cb5cbc49-vqj98" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.059749 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-6f96f8c84-xphlj" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.101240 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-8rkjv" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.130184 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-2knwl" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.156507 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-8ttcw" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.182088 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-kz58p" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.324577 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-ffks5" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.418550 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-sbzmh" Oct 09 08:33:05 crc kubenswrapper[4872]: I1009 08:33:05.461072 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-w6jlm" Oct 09 08:33:06 crc kubenswrapper[4872]: I1009 08:33:06.475039 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-585fc5b659-kgzkf" Oct 09 08:33:06 crc kubenswrapper[4872]: I1009 08:33:06.891873 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc" Oct 09 08:33:08 crc kubenswrapper[4872]: I1009 08:33:08.157528 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:33:08 crc kubenswrapper[4872]: I1009 08:33:08.157623 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:33:08 crc kubenswrapper[4872]: I1009 08:33:08.157717 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:33:08 crc kubenswrapper[4872]: I1009 08:33:08.158776 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:33:08 crc kubenswrapper[4872]: I1009 08:33:08.158880 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41" gracePeriod=600 Oct 09 08:33:09 crc kubenswrapper[4872]: I1009 08:33:09.481610 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41" exitCode=0 Oct 09 08:33:09 crc kubenswrapper[4872]: I1009 08:33:09.481749 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41"} Oct 09 08:33:09 crc kubenswrapper[4872]: I1009 08:33:09.482016 4872 scope.go:117] "RemoveContainer" containerID="c9656289297ac7f9ea947163781038a2a8266721c555f285a36d2d873443e5dd" Oct 09 08:33:12 crc kubenswrapper[4872]: I1009 08:33:12.508493 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" event={"ID":"a9091168-5676-4cd3-9a0d-f76d0c7c88df","Type":"ContainerStarted","Data":"090c6b80ccc335963a941ef92821cc4bdf04060404ce713763df6c2f6a3dd7d5"} Oct 09 08:33:12 crc kubenswrapper[4872]: I1009 08:33:12.509321 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:33:12 crc kubenswrapper[4872]: I1009 08:33:12.513878 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451"} Oct 09 08:33:12 crc kubenswrapper[4872]: I1009 08:33:12.530173 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" podStartSLOduration=3.051178871 podStartE2EDuration="28.530157557s" podCreationTimestamp="2025-10-09 08:32:44 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.389842493 +0000 UTC m=+924.580371119" lastFinishedPulling="2025-10-09 08:33:11.868821169 +0000 UTC m=+950.059349805" observedRunningTime="2025-10-09 08:33:12.526468011 +0000 UTC m=+950.716996637" watchObservedRunningTime="2025-10-09 08:33:12.530157557 +0000 UTC m=+950.720686183" Oct 09 08:33:13 crc kubenswrapper[4872]: I1009 08:33:13.522702 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" event={"ID":"add3400b-1fe2-4b5e-a810-7cca5a38ac96","Type":"ContainerStarted","Data":"abde22c1f77f5268613781dd85d96d021137e169d00c354e6e592be081ec1320"} Oct 09 08:33:13 crc kubenswrapper[4872]: I1009 08:33:13.545846 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz" podStartSLOduration=2.04174158 podStartE2EDuration="28.545828429s" podCreationTimestamp="2025-10-09 08:32:45 +0000 UTC" firstStartedPulling="2025-10-09 08:32:46.504674437 +0000 UTC m=+924.695203063" lastFinishedPulling="2025-10-09 08:33:13.008761286 +0000 UTC m=+951.199289912" observedRunningTime="2025-10-09 08:33:13.540166646 +0000 UTC m=+951.730695282" watchObservedRunningTime="2025-10-09 08:33:13.545828429 +0000 UTC m=+951.736357055" Oct 09 08:33:15 crc kubenswrapper[4872]: I1009 08:33:15.496285 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-ppxzv" Oct 09 08:33:15 crc kubenswrapper[4872]: I1009 08:33:15.516330 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-sjx2l" Oct 09 08:33:25 crc kubenswrapper[4872]: I1009 08:33:25.473857 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-b9zg2" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.807937 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.813153 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.818044 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.862722 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.862926 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.862963 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-gj269" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.862934 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.907438 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.909100 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.910922 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.919689 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.964890 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qrl5\" (UniqueName: \"kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:41 crc kubenswrapper[4872]: I1009 08:33:41.964966 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.066438 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qrl5\" (UniqueName: \"kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.066730 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.066816 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.066943 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.067036 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqn6z\" (UniqueName: \"kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.067722 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.089635 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qrl5\" (UniqueName: \"kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5\") pod \"dnsmasq-dns-675f4bcbfc-s8frx\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.168834 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.168984 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqn6z\" (UniqueName: \"kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.169189 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.169978 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.170906 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.182838 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.187383 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqn6z\" (UniqueName: \"kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z\") pod \"dnsmasq-dns-78dd6ddcc-c5wxv\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.224142 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.607754 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.614502 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.689096 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:42 crc kubenswrapper[4872]: W1009 08:33:42.693833 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43c2e28c_8b87_418d_8e56_3d49af199ad0.slice/crio-f51a4e3711a820cd87d42cd07575a9ec4cbbbaa5749b3cff2918571193ea0652 WatchSource:0}: Error finding container f51a4e3711a820cd87d42cd07575a9ec4cbbbaa5749b3cff2918571193ea0652: Status 404 returned error can't find the container with id f51a4e3711a820cd87d42cd07575a9ec4cbbbaa5749b3cff2918571193ea0652 Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.786678 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" event={"ID":"21029240-77a9-4534-a30c-2c32bcd1e673","Type":"ContainerStarted","Data":"4ed2ef9f71bee580a8faa1aed79db97ff36461016ffc0f7f0069f9fba8919282"} Oct 09 08:33:42 crc kubenswrapper[4872]: I1009 08:33:42.788184 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" event={"ID":"43c2e28c-8b87-418d-8e56-3d49af199ad0","Type":"ContainerStarted","Data":"f51a4e3711a820cd87d42cd07575a9ec4cbbbaa5749b3cff2918571193ea0652"} Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.766368 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.793412 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.794753 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.802957 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.911178 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gfqq\" (UniqueName: \"kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.911229 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:44 crc kubenswrapper[4872]: I1009 08:33:44.911289 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.014761 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.014845 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.014906 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gfqq\" (UniqueName: \"kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.015862 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.016284 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.084498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gfqq\" (UniqueName: \"kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq\") pod \"dnsmasq-dns-666b6646f7-kgpq7\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.097012 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.130867 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.134373 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.135830 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.165610 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.217075 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.217433 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lckp\" (UniqueName: \"kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.217463 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.321340 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.321407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.321493 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lckp\" (UniqueName: \"kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.322564 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.322980 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.350828 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lckp\" (UniqueName: \"kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp\") pod \"dnsmasq-dns-57d769cc4f-zw7hv\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.456849 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.767386 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.949844 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.952278 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.954619 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.956443 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.957185 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.957372 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.957520 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.957618 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.957604 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-d6zqp" Oct 09 08:33:45 crc kubenswrapper[4872]: I1009 08:33:45.970124 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134088 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134213 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134233 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134271 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lv6m9\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134351 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134373 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134397 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.134420 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235712 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235762 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235818 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235858 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lv6m9\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235916 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235938 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.235963 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.236144 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.236223 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.236324 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.236347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.236745 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.237005 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.237284 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.238042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.239654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.242676 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.243435 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.244864 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.245877 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.257288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lv6m9\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.258617 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.284803 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.414710 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.416421 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.420087 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.420210 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hm27g" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.420363 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.420372 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.420459 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.423132 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.423395 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.430477 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540834 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2s5s\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540883 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540912 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540936 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540954 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.540997 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.541025 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.541084 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.541104 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.541165 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.541181 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.642899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643301 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643359 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643399 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2s5s\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643504 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643536 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643545 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643590 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643679 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.643701 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.645370 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.645614 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.646583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.646758 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.647196 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.665622 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.666326 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.666881 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.674528 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2s5s\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.679819 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.690035 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:46 crc kubenswrapper[4872]: I1009 08:33:46.745718 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:33:47 crc kubenswrapper[4872]: W1009 08:33:47.965169 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4238d76_044c_4fde_8d78_48173a827317.slice/crio-ea5aa290319166fa41535ae569f9c371f4be56403e08cab9ca03c09d60e3bde0 WatchSource:0}: Error finding container ea5aa290319166fa41535ae569f9c371f4be56403e08cab9ca03c09d60e3bde0: Status 404 returned error can't find the container with id ea5aa290319166fa41535ae569f9c371f4be56403e08cab9ca03c09d60e3bde0 Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.213882 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.216205 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.218394 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.218499 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.218824 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-lfnfw" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.219711 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.221138 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.225524 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.226907 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371524 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371597 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371616 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqg4r\" (UniqueName: \"kubernetes.io/projected/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kube-api-access-bqg4r\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371654 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371688 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-secrets\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371723 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371757 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-default\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kolla-config\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.371797 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473346 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473419 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqg4r\" (UniqueName: \"kubernetes.io/projected/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kube-api-access-bqg4r\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473444 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-secrets\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473827 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473860 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-default\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473878 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kolla-config\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473893 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.473953 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-generated\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.474297 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.475023 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.475241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-config-data-default\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.475410 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kolla-config\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.475855 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-operator-scripts\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.478248 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-secrets\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.479489 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.490714 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.493526 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.494371 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqg4r\" (UniqueName: \"kubernetes.io/projected/0932ae43-84f6-4a0f-bf6e-fccd2f31c236-kube-api-access-bqg4r\") pod \"openstack-galera-0\" (UID: \"0932ae43-84f6-4a0f-bf6e-fccd2f31c236\") " pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.547610 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 09 08:33:48 crc kubenswrapper[4872]: I1009 08:33:48.844588 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" event={"ID":"e4238d76-044c-4fde-8d78-48173a827317","Type":"ContainerStarted","Data":"ea5aa290319166fa41535ae569f9c371f4be56403e08cab9ca03c09d60e3bde0"} Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.011243 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.012763 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.014073 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-j6mbf" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.015083 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.015271 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.015509 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.024299 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083014 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083198 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083218 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083431 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083491 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqhqt\" (UniqueName: \"kubernetes.io/projected/373d202a-3ea1-4855-a4d5-938451b3c42d-kube-api-access-jqhqt\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083549 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.083572 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.192772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.192850 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.192923 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.192952 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.192990 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.193055 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.193073 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.193189 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.193224 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqhqt\" (UniqueName: \"kubernetes.io/projected/373d202a-3ea1-4855-a4d5-938451b3c42d-kube-api-access-jqhqt\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.194494 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.194722 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.195212 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.197374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.199904 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.200012 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.211768 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/373d202a-3ea1-4855-a4d5-938451b3c42d-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.218334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/373d202a-3ea1-4855-a4d5-938451b3c42d-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.223160 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.236933 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqhqt\" (UniqueName: \"kubernetes.io/projected/373d202a-3ea1-4855-a4d5-938451b3c42d-kube-api-access-jqhqt\") pod \"openstack-cell1-galera-0\" (UID: \"373d202a-3ea1-4855-a4d5-938451b3c42d\") " pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.347263 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.375573 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.376511 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.381462 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.381794 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.382032 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-4v8bn" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.387040 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.497385 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-config-data\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.497444 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.497504 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk2gv\" (UniqueName: \"kubernetes.io/projected/5c8202d4-bb37-4cad-a36c-547b55adf925-kube-api-access-tk2gv\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.497524 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.497732 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-kolla-config\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.599219 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.599315 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk2gv\" (UniqueName: \"kubernetes.io/projected/5c8202d4-bb37-4cad-a36c-547b55adf925-kube-api-access-tk2gv\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.599344 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.599406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-kolla-config\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.599461 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-config-data\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.600211 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-config-data\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.600682 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5c8202d4-bb37-4cad-a36c-547b55adf925-kolla-config\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.603164 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-combined-ca-bundle\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.603580 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/5c8202d4-bb37-4cad-a36c-547b55adf925-memcached-tls-certs\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.616151 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk2gv\" (UniqueName: \"kubernetes.io/projected/5c8202d4-bb37-4cad-a36c-547b55adf925-kube-api-access-tk2gv\") pod \"memcached-0\" (UID: \"5c8202d4-bb37-4cad-a36c-547b55adf925\") " pod="openstack/memcached-0" Oct 09 08:33:49 crc kubenswrapper[4872]: I1009 08:33:49.693433 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.265170 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.266716 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.268490 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-zdz8p" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.275339 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.338152 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvb6h\" (UniqueName: \"kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h\") pod \"kube-state-metrics-0\" (UID: \"1effc4cb-04ae-4ee7-8de4-76502c225942\") " pod="openstack/kube-state-metrics-0" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.439156 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvb6h\" (UniqueName: \"kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h\") pod \"kube-state-metrics-0\" (UID: \"1effc4cb-04ae-4ee7-8de4-76502c225942\") " pod="openstack/kube-state-metrics-0" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.458551 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvb6h\" (UniqueName: \"kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h\") pod \"kube-state-metrics-0\" (UID: \"1effc4cb-04ae-4ee7-8de4-76502c225942\") " pod="openstack/kube-state-metrics-0" Oct 09 08:33:51 crc kubenswrapper[4872]: I1009 08:33:51.581715 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.668428 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zk2nb"] Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.669871 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.671821 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.672137 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.678968 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zk2nb"] Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.679997 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-chqgm" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.689503 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mfxrr"] Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.691173 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.728834 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mfxrr"] Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.818675 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-etc-ovs\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819032 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e454e39-b253-4330-b5f7-7619c2f0f1c9-scripts\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-log-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819104 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-lib\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819141 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-scripts\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819179 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-log\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819300 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-ovn-controller-tls-certs\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819327 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdmmq\" (UniqueName: \"kubernetes.io/projected/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-kube-api-access-jdmmq\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819346 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819409 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-combined-ca-bundle\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-run\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.819550 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxfjw\" (UniqueName: \"kubernetes.io/projected/2e454e39-b253-4330-b5f7-7619c2f0f1c9-kube-api-access-zxfjw\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920464 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-log-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920564 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-lib\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-scripts\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920612 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920694 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-log\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920712 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-ovn-controller-tls-certs\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920731 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdmmq\" (UniqueName: \"kubernetes.io/projected/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-kube-api-access-jdmmq\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920766 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920816 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-combined-ca-bundle\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920866 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-run\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920924 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxfjw\" (UniqueName: \"kubernetes.io/projected/2e454e39-b253-4330-b5f7-7619c2f0f1c9-kube-api-access-zxfjw\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920956 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e454e39-b253-4330-b5f7-7619c2f0f1c9-scripts\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.920972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-etc-ovs\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.921491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-etc-ovs\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.921761 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-log-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.921873 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-lib\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.923494 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-scripts\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.923605 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.923707 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-log\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.924617 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2e454e39-b253-4330-b5f7-7619c2f0f1c9-var-run-ovn\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.925537 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-var-run\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.926459 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2e454e39-b253-4330-b5f7-7619c2f0f1c9-scripts\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.931565 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-ovn-controller-tls-certs\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.935743 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e454e39-b253-4330-b5f7-7619c2f0f1c9-combined-ca-bundle\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.940271 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdmmq\" (UniqueName: \"kubernetes.io/projected/2a0f279e-d3c4-4d1a-b519-996c26f5ffba-kube-api-access-jdmmq\") pod \"ovn-controller-ovs-mfxrr\" (UID: \"2a0f279e-d3c4-4d1a-b519-996c26f5ffba\") " pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:55 crc kubenswrapper[4872]: I1009 08:33:55.945339 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxfjw\" (UniqueName: \"kubernetes.io/projected/2e454e39-b253-4330-b5f7-7619c2f0f1c9-kube-api-access-zxfjw\") pod \"ovn-controller-zk2nb\" (UID: \"2e454e39-b253-4330-b5f7-7619c2f0f1c9\") " pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:56 crc kubenswrapper[4872]: I1009 08:33:56.033398 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb" Oct 09 08:33:56 crc kubenswrapper[4872]: I1009 08:33:56.041963 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.648061 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.648223 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pqn6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-c5wxv_openstack(43c2e28c-8b87-418d-8e56-3d49af199ad0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.649456 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" podUID="43c2e28c-8b87-418d-8e56-3d49af199ad0" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.709919 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.710374 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8qrl5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-s8frx_openstack(21029240-77a9-4534-a30c-2c32bcd1e673): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 08:33:56 crc kubenswrapper[4872]: E1009 08:33:56.712213 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" podUID="21029240-77a9-4534-a30c-2c32bcd1e673" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.474703 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.481237 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.551254 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qrl5\" (UniqueName: \"kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5\") pod \"21029240-77a9-4534-a30c-2c32bcd1e673\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.551305 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc\") pod \"43c2e28c-8b87-418d-8e56-3d49af199ad0\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.551400 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config\") pod \"21029240-77a9-4534-a30c-2c32bcd1e673\" (UID: \"21029240-77a9-4534-a30c-2c32bcd1e673\") " Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.551429 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqn6z\" (UniqueName: \"kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z\") pod \"43c2e28c-8b87-418d-8e56-3d49af199ad0\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.551457 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config\") pod \"43c2e28c-8b87-418d-8e56-3d49af199ad0\" (UID: \"43c2e28c-8b87-418d-8e56-3d49af199ad0\") " Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.552525 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config" (OuterVolumeSpecName: "config") pod "43c2e28c-8b87-418d-8e56-3d49af199ad0" (UID: "43c2e28c-8b87-418d-8e56-3d49af199ad0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.554013 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43c2e28c-8b87-418d-8e56-3d49af199ad0" (UID: "43c2e28c-8b87-418d-8e56-3d49af199ad0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.554197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config" (OuterVolumeSpecName: "config") pod "21029240-77a9-4534-a30c-2c32bcd1e673" (UID: "21029240-77a9-4534-a30c-2c32bcd1e673"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.559538 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z" (OuterVolumeSpecName: "kube-api-access-pqn6z") pod "43c2e28c-8b87-418d-8e56-3d49af199ad0" (UID: "43c2e28c-8b87-418d-8e56-3d49af199ad0"). InnerVolumeSpecName "kube-api-access-pqn6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.560284 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5" (OuterVolumeSpecName: "kube-api-access-8qrl5") pod "21029240-77a9-4534-a30c-2c32bcd1e673" (UID: "21029240-77a9-4534-a30c-2c32bcd1e673"). InnerVolumeSpecName "kube-api-access-8qrl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.587331 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.603839 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.616265 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.626459 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.652911 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qrl5\" (UniqueName: \"kubernetes.io/projected/21029240-77a9-4534-a30c-2c32bcd1e673-kube-api-access-8qrl5\") on node \"crc\" DevicePath \"\"" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.652942 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.652951 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/21029240-77a9-4534-a30c-2c32bcd1e673-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.652960 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqn6z\" (UniqueName: \"kubernetes.io/projected/43c2e28c-8b87-418d-8e56-3d49af199ad0-kube-api-access-pqn6z\") on node \"crc\" DevicePath \"\"" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.652968 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43c2e28c-8b87-418d-8e56-3d49af199ad0-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.817505 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zk2nb"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.827722 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.878249 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.888295 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.926722 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerStarted","Data":"2ff0970f91397e74616bad4cd083744ea93b85359ba2daed45c8b16cecdb813f"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.928505 4872 generic.go:334] "Generic (PLEG): container finished" podID="e4238d76-044c-4fde-8d78-48173a827317" containerID="87a3a4977daf065d0108dffd7ae10bbf896e222d7f566ab5e539bf9314d2eef9" exitCode=0 Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.928622 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" event={"ID":"e4238d76-044c-4fde-8d78-48173a827317","Type":"ContainerDied","Data":"87a3a4977daf065d0108dffd7ae10bbf896e222d7f566ab5e539bf9314d2eef9"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.930440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" event={"ID":"21029240-77a9-4534-a30c-2c32bcd1e673","Type":"ContainerDied","Data":"4ed2ef9f71bee580a8faa1aed79db97ff36461016ffc0f7f0069f9fba8919282"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.930499 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8frx" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.935717 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.935717 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-c5wxv" event={"ID":"43c2e28c-8b87-418d-8e56-3d49af199ad0","Type":"ContainerDied","Data":"f51a4e3711a820cd87d42cd07575a9ec4cbbbaa5749b3cff2918571193ea0652"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.937639 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1effc4cb-04ae-4ee7-8de4-76502c225942","Type":"ContainerStarted","Data":"adaac7315338e7885def7de1a278a6691cc5e03eeb39345b95f09ebb4bf9698c"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.946993 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0932ae43-84f6-4a0f-bf6e-fccd2f31c236","Type":"ContainerStarted","Data":"456d99fead4cdb9e52ea69a7e491bd45c4667871cd646656788d99539e53654b"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.949563 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5c8202d4-bb37-4cad-a36c-547b55adf925","Type":"ContainerStarted","Data":"e35f702aebc3ff3daf8693f1c18efae81617ff8756b388450a2f49cb6d9c66c2"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.973621 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerStarted","Data":"67fe2f33135b08059728d92a02a8bc349605221425aa5ca3c7aed92eb39f8969"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.979736 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"373d202a-3ea1-4855-a4d5-938451b3c42d","Type":"ContainerStarted","Data":"83961a4e8e438a45bd9596c7afb2baefbf49fc0a373bd68cb3d5248bd19a3657"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.988175 4872 generic.go:334] "Generic (PLEG): container finished" podID="19d70327-8400-4925-b1ff-a4809607999c" containerID="38f1236806b6c0646509ea887ee4d8bd9fd63e55133295b4e2f313362ffb2d5f" exitCode=0 Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.988304 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" event={"ID":"19d70327-8400-4925-b1ff-a4809607999c","Type":"ContainerDied","Data":"38f1236806b6c0646509ea887ee4d8bd9fd63e55133295b4e2f313362ffb2d5f"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.988335 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" event={"ID":"19d70327-8400-4925-b1ff-a4809607999c","Type":"ContainerStarted","Data":"0b71249e596c855b087e4c21551f697fc3067c54b8c361c90ee015815faaa4ff"} Oct 09 08:33:57 crc kubenswrapper[4872]: I1009 08:33:57.991077 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zk2nb" event={"ID":"2e454e39-b253-4330-b5f7-7619c2f0f1c9","Type":"ContainerStarted","Data":"cd0790e667bcc4137e06e7a284ec758bc49adb456f9a373b3e42a75fa404baf7"} Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.045971 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mfxrr"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.079536 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.084799 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8frx"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.106207 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.129261 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-c5wxv"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.137486 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.146152 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.149452 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-72brr" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.149689 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.149800 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.149858 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.149956 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.157222 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268662 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268721 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268760 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268821 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268889 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268910 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hd28z\" (UniqueName: \"kubernetes.io/projected/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-kube-api-access-hd28z\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268959 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.268994 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.293065 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.294597 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.296846 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-qs5c5" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.296848 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.296859 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.297947 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.300960 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.370732 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.370788 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.370833 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.370928 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371000 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371035 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371095 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371166 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hd28z\" (UniqueName: \"kubernetes.io/projected/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-kube-api-access-hd28z\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371283 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371371 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371424 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371462 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npqsd\" (UniqueName: \"kubernetes.io/projected/81338d5d-c3ed-4163-90c3-0e60e578e35b-kube-api-access-npqsd\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371494 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371540 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.371557 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-config\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.372150 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.372899 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.373154 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.373312 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-config\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.377240 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.378082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.379144 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.399108 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hd28z\" (UniqueName: \"kubernetes.io/projected/f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a-kube-api-access-hd28z\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.410144 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a\") " pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.475976 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npqsd\" (UniqueName: \"kubernetes.io/projected/81338d5d-c3ed-4163-90c3-0e60e578e35b-kube-api-access-npqsd\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476074 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-config\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476158 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476187 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476290 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476334 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.476378 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.479566 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.480228 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.480475 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.481659 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-config\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.482499 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.483922 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.485714 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/81338d5d-c3ed-4163-90c3-0e60e578e35b-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.500785 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npqsd\" (UniqueName: \"kubernetes.io/projected/81338d5d-c3ed-4163-90c3-0e60e578e35b-kube-api-access-npqsd\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.502017 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21029240-77a9-4534-a30c-2c32bcd1e673" path="/var/lib/kubelet/pods/21029240-77a9-4534-a30c-2c32bcd1e673/volumes" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.506376 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43c2e28c-8b87-418d-8e56-3d49af199ad0" path="/var/lib/kubelet/pods/43c2e28c-8b87-418d-8e56-3d49af199ad0/volumes" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.506982 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/81338d5d-c3ed-4163-90c3-0e60e578e35b-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.517990 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"81338d5d-c3ed-4163-90c3-0e60e578e35b\") " pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:58 crc kubenswrapper[4872]: I1009 08:33:58.622097 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.013745 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" event={"ID":"19d70327-8400-4925-b1ff-a4809607999c","Type":"ContainerStarted","Data":"02df0d95b0663dcd5887e4e614ed1ecb047319613d510e4b46fd750406211d25"} Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.014378 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.017422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mfxrr" event={"ID":"2a0f279e-d3c4-4d1a-b519-996c26f5ffba","Type":"ContainerStarted","Data":"8cba95fa2b5b0c5411b964ecded5d3f3891c13399418b8deb3eec11c03347bc7"} Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.022777 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" event={"ID":"e4238d76-044c-4fde-8d78-48173a827317","Type":"ContainerStarted","Data":"d82e66ce028097cdbc63e6736f8ef236b3f2fecd5b61f8bc83516a3cd75a5302"} Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.023048 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.039253 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" podStartSLOduration=14.039025385 podStartE2EDuration="14.039025385s" podCreationTimestamp="2025-10-09 08:33:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:33:59.037993955 +0000 UTC m=+997.228522581" watchObservedRunningTime="2025-10-09 08:33:59.039025385 +0000 UTC m=+997.229554031" Oct 09 08:33:59 crc kubenswrapper[4872]: I1009 08:33:59.062374 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" podStartSLOduration=6.298129398 podStartE2EDuration="15.062341956s" podCreationTimestamp="2025-10-09 08:33:44 +0000 UTC" firstStartedPulling="2025-10-09 08:33:47.9778765 +0000 UTC m=+986.168405146" lastFinishedPulling="2025-10-09 08:33:56.742089078 +0000 UTC m=+994.932617704" observedRunningTime="2025-10-09 08:33:59.053236424 +0000 UTC m=+997.243765070" watchObservedRunningTime="2025-10-09 08:33:59.062341956 +0000 UTC m=+997.252870602" Oct 09 08:34:00 crc kubenswrapper[4872]: I1009 08:34:00.052601 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 09 08:34:00 crc kubenswrapper[4872]: I1009 08:34:00.199763 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 09 08:34:01 crc kubenswrapper[4872]: I1009 08:34:01.037102 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"81338d5d-c3ed-4163-90c3-0e60e578e35b","Type":"ContainerStarted","Data":"9357e8ace1bb50b5d3a195344e71e90c3b1ebee3ec7fbb8339515808eef77743"} Oct 09 08:34:01 crc kubenswrapper[4872]: W1009 08:34:01.710838 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf2fbc29a_eb3d_46bc_9bd8_10f4ce7b3c5a.slice/crio-4f7086c5f266b994da83205a3ccdfbb84c22f7613df19a221783e6fe7fa554a2 WatchSource:0}: Error finding container 4f7086c5f266b994da83205a3ccdfbb84c22f7613df19a221783e6fe7fa554a2: Status 404 returned error can't find the container with id 4f7086c5f266b994da83205a3ccdfbb84c22f7613df19a221783e6fe7fa554a2 Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.047196 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a","Type":"ContainerStarted","Data":"4f7086c5f266b994da83205a3ccdfbb84c22f7613df19a221783e6fe7fa554a2"} Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.259513 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-7p7x8"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.260913 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.262510 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.270443 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7p7x8"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346336 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-combined-ca-bundle\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346419 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prtcn\" (UniqueName: \"kubernetes.io/projected/19bf463a-82b1-4522-a8c7-0661281c1e50-kube-api-access-prtcn\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346447 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346534 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovs-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346572 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovn-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.346600 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bf463a-82b1-4522-a8c7-0661281c1e50-config\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.385574 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.385823 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="dnsmasq-dns" containerID="cri-o://d82e66ce028097cdbc63e6736f8ef236b3f2fecd5b61f8bc83516a3cd75a5302" gracePeriod=10 Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.417175 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.418484 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.420069 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.437390 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456024 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prtcn\" (UniqueName: \"kubernetes.io/projected/19bf463a-82b1-4522-a8c7-0661281c1e50-kube-api-access-prtcn\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456094 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456194 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovs-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456227 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovn-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456253 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bf463a-82b1-4522-a8c7-0661281c1e50-config\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.456309 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-combined-ca-bundle\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.457419 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovs-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.457830 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/19bf463a-82b1-4522-a8c7-0661281c1e50-ovn-rundir\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.463034 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19bf463a-82b1-4522-a8c7-0661281c1e50-config\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.468285 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-combined-ca-bundle\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.478446 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/19bf463a-82b1-4522-a8c7-0661281c1e50-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.495266 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prtcn\" (UniqueName: \"kubernetes.io/projected/19bf463a-82b1-4522-a8c7-0661281c1e50-kube-api-access-prtcn\") pod \"ovn-controller-metrics-7p7x8\" (UID: \"19bf463a-82b1-4522-a8c7-0661281c1e50\") " pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.560205 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.560271 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.560366 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.560386 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgcnw\" (UniqueName: \"kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.584532 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-7p7x8" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.596992 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.597218 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="dnsmasq-dns" containerID="cri-o://02df0d95b0663dcd5887e4e614ed1ecb047319613d510e4b46fd750406211d25" gracePeriod=10 Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.640711 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.642138 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.646897 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.648936 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.661753 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.661813 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.661899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.661926 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgcnw\" (UniqueName: \"kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.662894 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.663075 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.664334 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.686241 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgcnw\" (UniqueName: \"kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw\") pod \"dnsmasq-dns-5bf47b49b7-s28x4\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.735232 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.763320 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7b6lq\" (UniqueName: \"kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.763416 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.763461 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.763494 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.763562 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.865713 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7b6lq\" (UniqueName: \"kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.865807 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.865846 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.865881 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.865944 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.866927 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.866935 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.867039 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.867059 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.894167 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7b6lq\" (UniqueName: \"kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq\") pod \"dnsmasq-dns-8554648995-rzl2l\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:02 crc kubenswrapper[4872]: I1009 08:34:02.977730 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:03 crc kubenswrapper[4872]: I1009 08:34:03.057090 4872 generic.go:334] "Generic (PLEG): container finished" podID="e4238d76-044c-4fde-8d78-48173a827317" containerID="d82e66ce028097cdbc63e6736f8ef236b3f2fecd5b61f8bc83516a3cd75a5302" exitCode=0 Oct 09 08:34:03 crc kubenswrapper[4872]: I1009 08:34:03.057161 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" event={"ID":"e4238d76-044c-4fde-8d78-48173a827317","Type":"ContainerDied","Data":"d82e66ce028097cdbc63e6736f8ef236b3f2fecd5b61f8bc83516a3cd75a5302"} Oct 09 08:34:03 crc kubenswrapper[4872]: I1009 08:34:03.059711 4872 generic.go:334] "Generic (PLEG): container finished" podID="19d70327-8400-4925-b1ff-a4809607999c" containerID="02df0d95b0663dcd5887e4e614ed1ecb047319613d510e4b46fd750406211d25" exitCode=0 Oct 09 08:34:03 crc kubenswrapper[4872]: I1009 08:34:03.059756 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" event={"ID":"19d70327-8400-4925-b1ff-a4809607999c","Type":"ContainerDied","Data":"02df0d95b0663dcd5887e4e614ed1ecb047319613d510e4b46fd750406211d25"} Oct 09 08:34:05 crc kubenswrapper[4872]: I1009 08:34:05.134226 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.99:5353: connect: connection refused" Oct 09 08:34:05 crc kubenswrapper[4872]: I1009 08:34:05.458485 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.100:5353: connect: connection refused" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.202747 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.206856 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.326985 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lckp\" (UniqueName: \"kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp\") pod \"19d70327-8400-4925-b1ff-a4809607999c\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.327272 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gfqq\" (UniqueName: \"kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq\") pod \"e4238d76-044c-4fde-8d78-48173a827317\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.327292 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc\") pod \"19d70327-8400-4925-b1ff-a4809607999c\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.327334 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc\") pod \"e4238d76-044c-4fde-8d78-48173a827317\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.327355 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config\") pod \"e4238d76-044c-4fde-8d78-48173a827317\" (UID: \"e4238d76-044c-4fde-8d78-48173a827317\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.327468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config\") pod \"19d70327-8400-4925-b1ff-a4809607999c\" (UID: \"19d70327-8400-4925-b1ff-a4809607999c\") " Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.331047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp" (OuterVolumeSpecName: "kube-api-access-9lckp") pod "19d70327-8400-4925-b1ff-a4809607999c" (UID: "19d70327-8400-4925-b1ff-a4809607999c"). InnerVolumeSpecName "kube-api-access-9lckp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.341301 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq" (OuterVolumeSpecName: "kube-api-access-9gfqq") pod "e4238d76-044c-4fde-8d78-48173a827317" (UID: "e4238d76-044c-4fde-8d78-48173a827317"). InnerVolumeSpecName "kube-api-access-9gfqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.367323 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config" (OuterVolumeSpecName: "config") pod "19d70327-8400-4925-b1ff-a4809607999c" (UID: "19d70327-8400-4925-b1ff-a4809607999c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.367814 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e4238d76-044c-4fde-8d78-48173a827317" (UID: "e4238d76-044c-4fde-8d78-48173a827317"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.394145 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config" (OuterVolumeSpecName: "config") pod "e4238d76-044c-4fde-8d78-48173a827317" (UID: "e4238d76-044c-4fde-8d78-48173a827317"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.407474 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "19d70327-8400-4925-b1ff-a4809607999c" (UID: "19d70327-8400-4925-b1ff-a4809607999c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429062 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429299 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e4238d76-044c-4fde-8d78-48173a827317-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429309 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429318 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lckp\" (UniqueName: \"kubernetes.io/projected/19d70327-8400-4925-b1ff-a4809607999c-kube-api-access-9lckp\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429328 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gfqq\" (UniqueName: \"kubernetes.io/projected/e4238d76-044c-4fde-8d78-48173a827317-kube-api-access-9gfqq\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.429336 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/19d70327-8400-4925-b1ff-a4809607999c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.707069 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-7p7x8"] Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.798665 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:06 crc kubenswrapper[4872]: I1009 08:34:06.806018 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.089559 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" event={"ID":"e4238d76-044c-4fde-8d78-48173a827317","Type":"ContainerDied","Data":"ea5aa290319166fa41535ae569f9c371f4be56403e08cab9ca03c09d60e3bde0"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.089607 4872 scope.go:117] "RemoveContainer" containerID="d82e66ce028097cdbc63e6736f8ef236b3f2fecd5b61f8bc83516a3cd75a5302" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.089748 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-kgpq7" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.093035 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7p7x8" event={"ID":"19bf463a-82b1-4522-a8c7-0661281c1e50","Type":"ContainerStarted","Data":"782399eea40884d60298eb7f20a53bc198f44982febe811e5b3155c1c36e8cc8"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.095519 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0932ae43-84f6-4a0f-bf6e-fccd2f31c236","Type":"ContainerStarted","Data":"088edfb59bf631479d3d848ae9b088d2b418757f7b86197530c5d21eda134113"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.099499 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"81338d5d-c3ed-4163-90c3-0e60e578e35b","Type":"ContainerStarted","Data":"5c91895e72ff5e6e4920b952d0e10812b341b10bbd97f53870dcec7dd91e8a7f"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.102276 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" event={"ID":"19d70327-8400-4925-b1ff-a4809607999c","Type":"ContainerDied","Data":"0b71249e596c855b087e4c21551f697fc3067c54b8c361c90ee015815faaa4ff"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.102401 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zw7hv" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.106740 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" event={"ID":"cfc04f82-0342-4f8e-9589-a28793065e4d","Type":"ContainerStarted","Data":"0dbdc19ee356c07215c68ef7e3f5fdeabbd5ad0b3704692cd24151d283226e13"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.108919 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zk2nb" event={"ID":"2e454e39-b253-4330-b5f7-7619c2f0f1c9","Type":"ContainerStarted","Data":"b1baa5dcc0acda98af6008035ed6768ff688855eb1c7863890bd1e4ce44dd79a"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.108972 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-zk2nb" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.110549 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a","Type":"ContainerStarted","Data":"51cbce572574e01e2fd936060d6ea0e19861a54b4d453e7aa114adb5c26c2ab8"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.117937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"5c8202d4-bb37-4cad-a36c-547b55adf925","Type":"ContainerStarted","Data":"b0b71b9d6566bbb7298decc2ef245dcdf153b4473aa67aabd8fce1e61af0c9ce"} Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.118627 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.154365 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-zk2nb" podStartSLOduration=3.897310476 podStartE2EDuration="12.154348503s" podCreationTimestamp="2025-10-09 08:33:55 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.90129011 +0000 UTC m=+996.091818736" lastFinishedPulling="2025-10-09 08:34:06.158328117 +0000 UTC m=+1004.348856763" observedRunningTime="2025-10-09 08:34:07.15286355 +0000 UTC m=+1005.343392186" watchObservedRunningTime="2025-10-09 08:34:07.154348503 +0000 UTC m=+1005.344877129" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.174999 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=10.132567485 podStartE2EDuration="18.174950046s" podCreationTimestamp="2025-10-09 08:33:49 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.597975643 +0000 UTC m=+995.788504269" lastFinishedPulling="2025-10-09 08:34:05.640358204 +0000 UTC m=+1003.830886830" observedRunningTime="2025-10-09 08:34:07.167295005 +0000 UTC m=+1005.357823671" watchObservedRunningTime="2025-10-09 08:34:07.174950046 +0000 UTC m=+1005.365478682" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.661762 4872 scope.go:117] "RemoveContainer" containerID="87a3a4977daf065d0108dffd7ae10bbf896e222d7f566ab5e539bf9314d2eef9" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.679741 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.686435 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-kgpq7"] Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.703791 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.714193 4872 scope.go:117] "RemoveContainer" containerID="02df0d95b0663dcd5887e4e614ed1ecb047319613d510e4b46fd750406211d25" Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.714947 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zw7hv"] Oct 09 08:34:07 crc kubenswrapper[4872]: I1009 08:34:07.780267 4872 scope.go:117] "RemoveContainer" containerID="38f1236806b6c0646509ea887ee4d8bd9fd63e55133295b4e2f313362ffb2d5f" Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.129626 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"373d202a-3ea1-4855-a4d5-938451b3c42d","Type":"ContainerStarted","Data":"01696796a67bf3f0093a7bfbb11d944cbf149c6b7cb3c032b1bc3269f76dc95d"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.131818 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1effc4cb-04ae-4ee7-8de4-76502c225942","Type":"ContainerStarted","Data":"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.131914 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.134171 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerStarted","Data":"4d374843dc49fd7757ddcf08ee23c47d6bb3f10cabe4e1deb7a461f2b2087d12"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.136275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mfxrr" event={"ID":"2a0f279e-d3c4-4d1a-b519-996c26f5ffba","Type":"ContainerStarted","Data":"a750fcddedd3c8eb09237d4ab9facf835994176414d91adbc0223a3f6495f3b9"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.138077 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerStarted","Data":"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.142320 4872 generic.go:334] "Generic (PLEG): container finished" podID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerID="8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9" exitCode=0 Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.142390 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" event={"ID":"cfc04f82-0342-4f8e-9589-a28793065e4d","Type":"ContainerDied","Data":"8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.173257 4872 generic.go:334] "Generic (PLEG): container finished" podID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerID="05913476fe992fb02ed6e1b39cf2e9a7c4022036991372b21d61b4fc436f6ff4" exitCode=0 Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.173736 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rzl2l" event={"ID":"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d","Type":"ContainerDied","Data":"05913476fe992fb02ed6e1b39cf2e9a7c4022036991372b21d61b4fc436f6ff4"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.173829 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rzl2l" event={"ID":"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d","Type":"ContainerStarted","Data":"052e1b30f37ea3199c13f9a3445bab20a5555e3acc6eb3694305efa07ece0c32"} Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.209785 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.596495393 podStartE2EDuration="17.209767529s" podCreationTimestamp="2025-10-09 08:33:51 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.617770593 +0000 UTC m=+995.808299219" lastFinishedPulling="2025-10-09 08:34:06.231042729 +0000 UTC m=+1004.421571355" observedRunningTime="2025-10-09 08:34:08.207182415 +0000 UTC m=+1006.397711061" watchObservedRunningTime="2025-10-09 08:34:08.209767529 +0000 UTC m=+1006.400296155" Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.473991 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19d70327-8400-4925-b1ff-a4809607999c" path="/var/lib/kubelet/pods/19d70327-8400-4925-b1ff-a4809607999c/volumes" Oct 09 08:34:08 crc kubenswrapper[4872]: I1009 08:34:08.474746 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4238d76-044c-4fde-8d78-48173a827317" path="/var/lib/kubelet/pods/e4238d76-044c-4fde-8d78-48173a827317/volumes" Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.183508 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a0f279e-d3c4-4d1a-b519-996c26f5ffba" containerID="a750fcddedd3c8eb09237d4ab9facf835994176414d91adbc0223a3f6495f3b9" exitCode=0 Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.183683 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mfxrr" event={"ID":"2a0f279e-d3c4-4d1a-b519-996c26f5ffba","Type":"ContainerDied","Data":"a750fcddedd3c8eb09237d4ab9facf835994176414d91adbc0223a3f6495f3b9"} Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.200936 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rzl2l" event={"ID":"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d","Type":"ContainerStarted","Data":"efa98f4092add0671dce379c54c76eabb5c3de4a3658ade5606e511516f430a1"} Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.201408 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.203850 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" event={"ID":"cfc04f82-0342-4f8e-9589-a28793065e4d","Type":"ContainerStarted","Data":"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b"} Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.203880 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.239164 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" podStartSLOduration=7.239142756 podStartE2EDuration="7.239142756s" podCreationTimestamp="2025-10-09 08:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:34:09.234501692 +0000 UTC m=+1007.425030338" watchObservedRunningTime="2025-10-09 08:34:09.239142756 +0000 UTC m=+1007.429671382" Oct 09 08:34:09 crc kubenswrapper[4872]: I1009 08:34:09.258407 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-rzl2l" podStartSLOduration=7.258388289 podStartE2EDuration="7.258388289s" podCreationTimestamp="2025-10-09 08:34:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:34:09.25111903 +0000 UTC m=+1007.441647676" watchObservedRunningTime="2025-10-09 08:34:09.258388289 +0000 UTC m=+1007.448916915" Oct 09 08:34:11 crc kubenswrapper[4872]: I1009 08:34:11.217933 4872 generic.go:334] "Generic (PLEG): container finished" podID="0932ae43-84f6-4a0f-bf6e-fccd2f31c236" containerID="088edfb59bf631479d3d848ae9b088d2b418757f7b86197530c5d21eda134113" exitCode=0 Oct 09 08:34:11 crc kubenswrapper[4872]: I1009 08:34:11.218013 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0932ae43-84f6-4a0f-bf6e-fccd2f31c236","Type":"ContainerDied","Data":"088edfb59bf631479d3d848ae9b088d2b418757f7b86197530c5d21eda134113"} Oct 09 08:34:14 crc kubenswrapper[4872]: I1009 08:34:14.695589 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 09 08:34:17 crc kubenswrapper[4872]: I1009 08:34:17.736786 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:17 crc kubenswrapper[4872]: I1009 08:34:17.980835 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.037614 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.272057 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mfxrr" event={"ID":"2a0f279e-d3c4-4d1a-b519-996c26f5ffba","Type":"ContainerStarted","Data":"85104b2a260e5f8126bf66e0384ebc0939f000ac730f2728a1903d9ffaa9428a"} Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.273397 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-7p7x8" event={"ID":"19bf463a-82b1-4522-a8c7-0661281c1e50","Type":"ContainerStarted","Data":"3191481a0cfcac5f33a3a3b88679fa2107160ca2ec77c85e698f4dc189c45afd"} Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.275249 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a","Type":"ContainerStarted","Data":"4b3cbc3a27fffdd9699f31c9903c139a642f95745da1bdc215fba01fc743aaed"} Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.277122 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="dnsmasq-dns" containerID="cri-o://b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b" gracePeriod=10 Oct 09 08:34:18 crc kubenswrapper[4872]: I1009 08:34:18.277437 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"81338d5d-c3ed-4163-90c3-0e60e578e35b","Type":"ContainerStarted","Data":"e9fb78b899e4e5e2ce89aae0e52001ecc2765e44a1b311471ec629d46aa47d44"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.199967 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.287895 4872 generic.go:334] "Generic (PLEG): container finished" podID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerID="b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b" exitCode=0 Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.287963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" event={"ID":"cfc04f82-0342-4f8e-9589-a28793065e4d","Type":"ContainerDied","Data":"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.287991 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" event={"ID":"cfc04f82-0342-4f8e-9589-a28793065e4d","Type":"ContainerDied","Data":"0dbdc19ee356c07215c68ef7e3f5fdeabbd5ad0b3704692cd24151d283226e13"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.288008 4872 scope.go:117] "RemoveContainer" containerID="b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.288010 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-s28x4" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.290936 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mfxrr" event={"ID":"2a0f279e-d3c4-4d1a-b519-996c26f5ffba","Type":"ContainerStarted","Data":"dd28613f65c3a0d21d196a16aa5d87fc8ef76ca28bfe96497577b8cb7005b06b"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.291016 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.291137 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.295299 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"0932ae43-84f6-4a0f-bf6e-fccd2f31c236","Type":"ContainerStarted","Data":"c5ee437221305cffc4e60b5d7b6d568c9f429953b0bafa6b3ab99a45e3482c17"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.298673 4872 generic.go:334] "Generic (PLEG): container finished" podID="373d202a-3ea1-4855-a4d5-938451b3c42d" containerID="01696796a67bf3f0093a7bfbb11d944cbf149c6b7cb3c032b1bc3269f76dc95d" exitCode=0 Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.298751 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"373d202a-3ea1-4855-a4d5-938451b3c42d","Type":"ContainerDied","Data":"01696796a67bf3f0093a7bfbb11d944cbf149c6b7cb3c032b1bc3269f76dc95d"} Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.330211 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mfxrr" podStartSLOduration=16.156973265 podStartE2EDuration="24.33019216s" podCreationTimestamp="2025-10-09 08:33:55 +0000 UTC" firstStartedPulling="2025-10-09 08:33:58.057910096 +0000 UTC m=+996.248438722" lastFinishedPulling="2025-10-09 08:34:06.231128991 +0000 UTC m=+1004.421657617" observedRunningTime="2025-10-09 08:34:19.321906691 +0000 UTC m=+1017.512435337" watchObservedRunningTime="2025-10-09 08:34:19.33019216 +0000 UTC m=+1017.520720786" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.343145 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-7p7x8" podStartSLOduration=14.009589422 podStartE2EDuration="17.343124282s" podCreationTimestamp="2025-10-09 08:34:02 +0000 UTC" firstStartedPulling="2025-10-09 08:34:06.885891139 +0000 UTC m=+1005.076419775" lastFinishedPulling="2025-10-09 08:34:10.219426009 +0000 UTC m=+1008.409954635" observedRunningTime="2025-10-09 08:34:19.339813247 +0000 UTC m=+1017.530341893" watchObservedRunningTime="2025-10-09 08:34:19.343124282 +0000 UTC m=+1017.533652928" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.343817 4872 scope.go:117] "RemoveContainer" containerID="8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.358367 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgcnw\" (UniqueName: \"kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw\") pod \"cfc04f82-0342-4f8e-9589-a28793065e4d\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.358466 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config\") pod \"cfc04f82-0342-4f8e-9589-a28793065e4d\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.358515 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb\") pod \"cfc04f82-0342-4f8e-9589-a28793065e4d\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.358585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc\") pod \"cfc04f82-0342-4f8e-9589-a28793065e4d\" (UID: \"cfc04f82-0342-4f8e-9589-a28793065e4d\") " Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.362783 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw" (OuterVolumeSpecName: "kube-api-access-rgcnw") pod "cfc04f82-0342-4f8e-9589-a28793065e4d" (UID: "cfc04f82-0342-4f8e-9589-a28793065e4d"). InnerVolumeSpecName "kube-api-access-rgcnw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.386808 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=24.134432996 podStartE2EDuration="32.386695015s" podCreationTimestamp="2025-10-09 08:33:47 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.905533092 +0000 UTC m=+996.096061718" lastFinishedPulling="2025-10-09 08:34:06.157795091 +0000 UTC m=+1004.348323737" observedRunningTime="2025-10-09 08:34:19.373362932 +0000 UTC m=+1017.563891578" watchObservedRunningTime="2025-10-09 08:34:19.386695015 +0000 UTC m=+1017.577223641" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.400903 4872 scope.go:117] "RemoveContainer" containerID="b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b" Oct 09 08:34:19 crc kubenswrapper[4872]: E1009 08:34:19.401424 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b\": container with ID starting with b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b not found: ID does not exist" containerID="b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.401482 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b"} err="failed to get container status \"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b\": rpc error: code = NotFound desc = could not find container \"b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b\": container with ID starting with b1164e6aa99d5e90e5ad28c1cffef855bb3ade7573f59624d1db4567e55df69b not found: ID does not exist" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.401513 4872 scope.go:117] "RemoveContainer" containerID="8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9" Oct 09 08:34:19 crc kubenswrapper[4872]: E1009 08:34:19.405230 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9\": container with ID starting with 8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9 not found: ID does not exist" containerID="8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.405277 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9"} err="failed to get container status \"8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9\": rpc error: code = NotFound desc = could not find container \"8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9\": container with ID starting with 8641bdf79140ff09d52ce93432cfbe00fbec6db478225346ba95d0300491eda9 not found: ID does not exist" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.412862 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=12.889416355 podStartE2EDuration="22.412834887s" podCreationTimestamp="2025-10-09 08:33:57 +0000 UTC" firstStartedPulling="2025-10-09 08:34:00.746947293 +0000 UTC m=+998.937475919" lastFinishedPulling="2025-10-09 08:34:10.270365825 +0000 UTC m=+1008.460894451" observedRunningTime="2025-10-09 08:34:19.399318659 +0000 UTC m=+1017.589847295" watchObservedRunningTime="2025-10-09 08:34:19.412834887 +0000 UTC m=+1017.603363513" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.420799 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cfc04f82-0342-4f8e-9589-a28793065e4d" (UID: "cfc04f82-0342-4f8e-9589-a28793065e4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.423935 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config" (OuterVolumeSpecName: "config") pod "cfc04f82-0342-4f8e-9589-a28793065e4d" (UID: "cfc04f82-0342-4f8e-9589-a28793065e4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.454542 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cfc04f82-0342-4f8e-9589-a28793065e4d" (UID: "cfc04f82-0342-4f8e-9589-a28793065e4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.462247 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgcnw\" (UniqueName: \"kubernetes.io/projected/cfc04f82-0342-4f8e-9589-a28793065e4d-kube-api-access-rgcnw\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.462297 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.462321 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.462334 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cfc04f82-0342-4f8e-9589-a28793065e4d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.481505 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.488473 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=13.955684975 podStartE2EDuration="22.488449763s" podCreationTimestamp="2025-10-09 08:33:57 +0000 UTC" firstStartedPulling="2025-10-09 08:34:01.713185264 +0000 UTC m=+999.903713890" lastFinishedPulling="2025-10-09 08:34:10.245950032 +0000 UTC m=+1008.436478678" observedRunningTime="2025-10-09 08:34:19.484461108 +0000 UTC m=+1017.674989754" watchObservedRunningTime="2025-10-09 08:34:19.488449763 +0000 UTC m=+1017.678978389" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.527631 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.617117 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.623777 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.623811 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-s28x4"] Oct 09 08:34:19 crc kubenswrapper[4872]: I1009 08:34:19.667016 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.314009 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"373d202a-3ea1-4855-a4d5-938451b3c42d","Type":"ContainerStarted","Data":"4de23234c2f59865fdb26b86d6c47664327799648dbe74e329c9ae436c2470ca"} Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.316597 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.317066 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.349195 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=24.965795516 podStartE2EDuration="33.349167267s" podCreationTimestamp="2025-10-09 08:33:47 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.90545408 +0000 UTC m=+996.095982706" lastFinishedPulling="2025-10-09 08:34:06.288825831 +0000 UTC m=+1004.479354457" observedRunningTime="2025-10-09 08:34:20.345088 +0000 UTC m=+1018.535616656" watchObservedRunningTime="2025-10-09 08:34:20.349167267 +0000 UTC m=+1018.539695933" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.371301 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.381281 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.472349 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" path="/var/lib/kubelet/pods/cfc04f82-0342-4f8e-9589-a28793065e4d/volumes" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622201 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622534 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622558 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622584 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622591 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622602 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622608 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622624 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622631 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622663 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622669 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: E1009 08:34:20.622687 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622694 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="init" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622855 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="19d70327-8400-4925-b1ff-a4809607999c" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622878 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4238d76-044c-4fde-8d78-48173a827317" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.622888 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfc04f82-0342-4f8e-9589-a28793065e4d" containerName="dnsmasq-dns" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.623862 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.634512 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.634704 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.634712 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-c8d6f" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.643702 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.643948 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.683875 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684115 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-config\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684208 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-scripts\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684325 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684425 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.684660 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljph8\" (UniqueName: \"kubernetes.io/projected/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-kube-api-access-ljph8\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786510 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786568 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-config\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786599 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-scripts\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786654 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786699 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786719 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.786746 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljph8\" (UniqueName: \"kubernetes.io/projected/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-kube-api-access-ljph8\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.787797 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.787857 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-config\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.788388 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-scripts\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.792510 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.796776 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.801405 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.811322 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljph8\" (UniqueName: \"kubernetes.io/projected/4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85-kube-api-access-ljph8\") pod \"ovn-northd-0\" (UID: \"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85\") " pod="openstack/ovn-northd-0" Oct 09 08:34:20 crc kubenswrapper[4872]: I1009 08:34:20.941687 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.394351 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 09 08:34:21 crc kubenswrapper[4872]: W1009 08:34:21.400830 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f7b58e0_4eea_4ec2_b3e3_df0ccdf9aa85.slice/crio-335bfd694abe57cd31536330eb29e12a9958cef07793a8b11bcb9ee2111e90db WatchSource:0}: Error finding container 335bfd694abe57cd31536330eb29e12a9958cef07793a8b11bcb9ee2111e90db: Status 404 returned error can't find the container with id 335bfd694abe57cd31536330eb29e12a9958cef07793a8b11bcb9ee2111e90db Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.595295 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.613013 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.614209 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.641391 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.702147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.702205 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5h66p\" (UniqueName: \"kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.702271 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.702347 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.702418 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.811055 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.811216 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.811258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5h66p\" (UniqueName: \"kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.811307 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.811433 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.813150 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.813225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.813808 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.817260 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.845616 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5h66p\" (UniqueName: \"kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p\") pod \"dnsmasq-dns-b8fbc5445-gl5sb\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:21 crc kubenswrapper[4872]: I1009 08:34:21.940395 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.333938 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85","Type":"ContainerStarted","Data":"335bfd694abe57cd31536330eb29e12a9958cef07793a8b11bcb9ee2111e90db"} Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.427436 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:22 crc kubenswrapper[4872]: W1009 08:34:22.625992 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podad0e49d0_0cc0_4f1d_8212_f798c25d4017.slice/crio-0e71c37dd14d299f3a264f4b109506c349658e9a5cec5ee261a4155da0f4a037 WatchSource:0}: Error finding container 0e71c37dd14d299f3a264f4b109506c349658e9a5cec5ee261a4155da0f4a037: Status 404 returned error can't find the container with id 0e71c37dd14d299f3a264f4b109506c349658e9a5cec5ee261a4155da0f4a037 Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.752206 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.760389 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.766854 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-5scqs" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.767099 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.767280 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.767430 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.778891 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.936114 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.936193 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-cache\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.936234 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.936288 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zmb7\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-kube-api-access-5zmb7\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:22 crc kubenswrapper[4872]: I1009 08:34:22.936348 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-lock\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.038512 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.038570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-cache\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.038610 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.038663 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zmb7\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-kube-api-access-5zmb7\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.038698 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-lock\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.039129 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-lock\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.039237 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.039254 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.039294 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift podName:6ee9bdf9-4380-47b9-b1b9-083974d8f01a nodeName:}" failed. No retries permitted until 2025-10-09 08:34:23.539276235 +0000 UTC m=+1021.729804861 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift") pod "swift-storage-0" (UID: "6ee9bdf9-4380-47b9-b1b9-083974d8f01a") : configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.039911 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-cache\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.039943 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.062612 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zmb7\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-kube-api-access-5zmb7\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.107092 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.309612 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-qx7q8"] Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.310572 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.313149 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.315243 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.318344 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.326129 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qx7q8"] Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.342463 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85","Type":"ContainerStarted","Data":"d07e68b6315c42e1f36d6e6343e38d043869f1ecbd3411454ab2eb5447aa3c13"} Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.342512 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85","Type":"ContainerStarted","Data":"8b4bcf06f876e930ccd36876790994c828b41183f72ce73b91d0ab25c448265c"} Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.342729 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.345856 4872 generic.go:334] "Generic (PLEG): container finished" podID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerID="3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad" exitCode=0 Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.345894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" event={"ID":"ad0e49d0-0cc0-4f1d-8212-f798c25d4017","Type":"ContainerDied","Data":"3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad"} Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.345964 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" event={"ID":"ad0e49d0-0cc0-4f1d-8212-f798c25d4017","Type":"ContainerStarted","Data":"0e71c37dd14d299f3a264f4b109506c349658e9a5cec5ee261a4155da0f4a037"} Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.346955 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.346995 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.347147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.347218 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.347283 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpl59\" (UniqueName: \"kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.347328 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.347379 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.401389 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.108633088 podStartE2EDuration="3.401373273s" podCreationTimestamp="2025-10-09 08:34:20 +0000 UTC" firstStartedPulling="2025-10-09 08:34:21.403013527 +0000 UTC m=+1019.593542153" lastFinishedPulling="2025-10-09 08:34:22.695753712 +0000 UTC m=+1020.886282338" observedRunningTime="2025-10-09 08:34:23.365697467 +0000 UTC m=+1021.556226103" watchObservedRunningTime="2025-10-09 08:34:23.401373273 +0000 UTC m=+1021.591901899" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448011 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448730 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448757 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448778 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpl59\" (UniqueName: \"kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448803 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.448838 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.449313 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.449473 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.449922 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.452049 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.453423 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.454421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.465230 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpl59\" (UniqueName: \"kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59\") pod \"swift-ring-rebalance-qx7q8\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.550617 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.550829 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.550848 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: E1009 08:34:23.550900 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift podName:6ee9bdf9-4380-47b9-b1b9-083974d8f01a nodeName:}" failed. No retries permitted until 2025-10-09 08:34:24.550885625 +0000 UTC m=+1022.741414251 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift") pod "swift-storage-0" (UID: "6ee9bdf9-4380-47b9-b1b9-083974d8f01a") : configmap "swift-ring-files" not found Oct 09 08:34:23 crc kubenswrapper[4872]: I1009 08:34:23.666286 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:24 crc kubenswrapper[4872]: I1009 08:34:24.115438 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-qx7q8"] Oct 09 08:34:24 crc kubenswrapper[4872]: W1009 08:34:24.129018 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31fd23b6_fa18_4681_ba21_b9e69ad958d0.slice/crio-a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4 WatchSource:0}: Error finding container a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4: Status 404 returned error can't find the container with id a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4 Oct 09 08:34:24 crc kubenswrapper[4872]: I1009 08:34:24.354427 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qx7q8" event={"ID":"31fd23b6-fa18-4681-ba21-b9e69ad958d0","Type":"ContainerStarted","Data":"a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4"} Oct 09 08:34:24 crc kubenswrapper[4872]: I1009 08:34:24.356498 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" event={"ID":"ad0e49d0-0cc0-4f1d-8212-f798c25d4017","Type":"ContainerStarted","Data":"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724"} Oct 09 08:34:24 crc kubenswrapper[4872]: I1009 08:34:24.564380 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:24 crc kubenswrapper[4872]: E1009 08:34:24.564899 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 09 08:34:24 crc kubenswrapper[4872]: E1009 08:34:24.564948 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 09 08:34:24 crc kubenswrapper[4872]: E1009 08:34:24.565007 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift podName:6ee9bdf9-4380-47b9-b1b9-083974d8f01a nodeName:}" failed. No retries permitted until 2025-10-09 08:34:26.564988021 +0000 UTC m=+1024.755516717 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift") pod "swift-storage-0" (UID: "6ee9bdf9-4380-47b9-b1b9-083974d8f01a") : configmap "swift-ring-files" not found Oct 09 08:34:25 crc kubenswrapper[4872]: I1009 08:34:25.365112 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:26 crc kubenswrapper[4872]: I1009 08:34:26.600729 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:26 crc kubenswrapper[4872]: E1009 08:34:26.600966 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 09 08:34:26 crc kubenswrapper[4872]: E1009 08:34:26.601184 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 09 08:34:26 crc kubenswrapper[4872]: E1009 08:34:26.601249 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift podName:6ee9bdf9-4380-47b9-b1b9-083974d8f01a nodeName:}" failed. No retries permitted until 2025-10-09 08:34:30.601231747 +0000 UTC m=+1028.791760373 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift") pod "swift-storage-0" (UID: "6ee9bdf9-4380-47b9-b1b9-083974d8f01a") : configmap "swift-ring-files" not found Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.390357 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qx7q8" event={"ID":"31fd23b6-fa18-4681-ba21-b9e69ad958d0","Type":"ContainerStarted","Data":"92b02df8324395bd0df06a21434ae07095a74d3fba44daa130caad1fa833a873"} Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.414102 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" podStartSLOduration=7.414082195 podStartE2EDuration="7.414082195s" podCreationTimestamp="2025-10-09 08:34:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:34:24.374753699 +0000 UTC m=+1022.565282345" watchObservedRunningTime="2025-10-09 08:34:28.414082195 +0000 UTC m=+1026.604610841" Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.549375 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.549460 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.593286 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 09 08:34:28 crc kubenswrapper[4872]: I1009 08:34:28.616478 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-qx7q8" podStartSLOduration=2.529243615 podStartE2EDuration="5.616462158s" podCreationTimestamp="2025-10-09 08:34:23 +0000 UTC" firstStartedPulling="2025-10-09 08:34:24.130716068 +0000 UTC m=+1022.321244704" lastFinishedPulling="2025-10-09 08:34:27.217934621 +0000 UTC m=+1025.408463247" observedRunningTime="2025-10-09 08:34:28.410600285 +0000 UTC m=+1026.601128921" watchObservedRunningTime="2025-10-09 08:34:28.616462158 +0000 UTC m=+1026.806990784" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.347952 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.348280 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.406711 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.455925 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.457729 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.700657 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nxr9k"] Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.701872 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.710012 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nxr9k"] Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.854422 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhnbz\" (UniqueName: \"kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz\") pod \"placement-db-create-nxr9k\" (UID: \"e189a93b-9f2f-4c99-b315-69473291ed9c\") " pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.955855 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhnbz\" (UniqueName: \"kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz\") pod \"placement-db-create-nxr9k\" (UID: \"e189a93b-9f2f-4c99-b315-69473291ed9c\") " pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:29 crc kubenswrapper[4872]: I1009 08:34:29.974516 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhnbz\" (UniqueName: \"kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz\") pod \"placement-db-create-nxr9k\" (UID: \"e189a93b-9f2f-4c99-b315-69473291ed9c\") " pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:30 crc kubenswrapper[4872]: I1009 08:34:30.017485 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:30 crc kubenswrapper[4872]: I1009 08:34:30.457526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nxr9k"] Oct 09 08:34:30 crc kubenswrapper[4872]: I1009 08:34:30.668478 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:30 crc kubenswrapper[4872]: E1009 08:34:30.668722 4872 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 09 08:34:30 crc kubenswrapper[4872]: E1009 08:34:30.669143 4872 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 09 08:34:30 crc kubenswrapper[4872]: E1009 08:34:30.669208 4872 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift podName:6ee9bdf9-4380-47b9-b1b9-083974d8f01a nodeName:}" failed. No retries permitted until 2025-10-09 08:34:38.669187948 +0000 UTC m=+1036.859716564 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift") pod "swift-storage-0" (UID: "6ee9bdf9-4380-47b9-b1b9-083974d8f01a") : configmap "swift-ring-files" not found Oct 09 08:34:31 crc kubenswrapper[4872]: I1009 08:34:31.414561 4872 generic.go:334] "Generic (PLEG): container finished" podID="e189a93b-9f2f-4c99-b315-69473291ed9c" containerID="f2a6ee52f21536fb1dc6a270480ccc5c38f6d1a4344d77078d462f26453869ae" exitCode=0 Oct 09 08:34:31 crc kubenswrapper[4872]: I1009 08:34:31.414623 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nxr9k" event={"ID":"e189a93b-9f2f-4c99-b315-69473291ed9c","Type":"ContainerDied","Data":"f2a6ee52f21536fb1dc6a270480ccc5c38f6d1a4344d77078d462f26453869ae"} Oct 09 08:34:31 crc kubenswrapper[4872]: I1009 08:34:31.414676 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nxr9k" event={"ID":"e189a93b-9f2f-4c99-b315-69473291ed9c","Type":"ContainerStarted","Data":"0adcf3ad9885ef9099bc71c00c31f52c00d5945812c197760b05756660935470"} Oct 09 08:34:31 crc kubenswrapper[4872]: I1009 08:34:31.942676 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.008409 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.008907 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-rzl2l" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="dnsmasq-dns" containerID="cri-o://efa98f4092add0671dce379c54c76eabb5c3de4a3658ade5606e511516f430a1" gracePeriod=10 Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.423324 4872 generic.go:334] "Generic (PLEG): container finished" podID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerID="efa98f4092add0671dce379c54c76eabb5c3de4a3658ade5606e511516f430a1" exitCode=0 Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.423369 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rzl2l" event={"ID":"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d","Type":"ContainerDied","Data":"efa98f4092add0671dce379c54c76eabb5c3de4a3658ade5606e511516f430a1"} Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.527967 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.707271 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb\") pod \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.707419 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb\") pod \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.707437 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc\") pod \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.707529 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config\") pod \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.707565 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7b6lq\" (UniqueName: \"kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq\") pod \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\" (UID: \"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.713277 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq" (OuterVolumeSpecName: "kube-api-access-7b6lq") pod "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" (UID: "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d"). InnerVolumeSpecName "kube-api-access-7b6lq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.729921 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.750449 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" (UID: "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.757939 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config" (OuterVolumeSpecName: "config") pod "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" (UID: "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.762323 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" (UID: "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.765759 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" (UID: "0a9f72e3-4a7d-4d44-a14f-f752a160ad3d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.810605 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7b6lq\" (UniqueName: \"kubernetes.io/projected/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-kube-api-access-7b6lq\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.810660 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.810674 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.810686 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.810703 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.913070 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhnbz\" (UniqueName: \"kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz\") pod \"e189a93b-9f2f-4c99-b315-69473291ed9c\" (UID: \"e189a93b-9f2f-4c99-b315-69473291ed9c\") " Oct 09 08:34:32 crc kubenswrapper[4872]: I1009 08:34:32.916797 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz" (OuterVolumeSpecName: "kube-api-access-vhnbz") pod "e189a93b-9f2f-4c99-b315-69473291ed9c" (UID: "e189a93b-9f2f-4c99-b315-69473291ed9c"). InnerVolumeSpecName "kube-api-access-vhnbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.014747 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhnbz\" (UniqueName: \"kubernetes.io/projected/e189a93b-9f2f-4c99-b315-69473291ed9c-kube-api-access-vhnbz\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.434457 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-rzl2l" event={"ID":"0a9f72e3-4a7d-4d44-a14f-f752a160ad3d","Type":"ContainerDied","Data":"052e1b30f37ea3199c13f9a3445bab20a5555e3acc6eb3694305efa07ece0c32"} Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.434519 4872 scope.go:117] "RemoveContainer" containerID="efa98f4092add0671dce379c54c76eabb5c3de4a3658ade5606e511516f430a1" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.434486 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-rzl2l" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.436294 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nxr9k" event={"ID":"e189a93b-9f2f-4c99-b315-69473291ed9c","Type":"ContainerDied","Data":"0adcf3ad9885ef9099bc71c00c31f52c00d5945812c197760b05756660935470"} Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.436337 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nxr9k" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.436344 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0adcf3ad9885ef9099bc71c00c31f52c00d5945812c197760b05756660935470" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.458837 4872 scope.go:117] "RemoveContainer" containerID="05913476fe992fb02ed6e1b39cf2e9a7c4022036991372b21d61b4fc436f6ff4" Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.498910 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:33 crc kubenswrapper[4872]: I1009 08:34:33.506695 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-rzl2l"] Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.448349 4872 generic.go:334] "Generic (PLEG): container finished" podID="31fd23b6-fa18-4681-ba21-b9e69ad958d0" containerID="92b02df8324395bd0df06a21434ae07095a74d3fba44daa130caad1fa833a873" exitCode=0 Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.448433 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qx7q8" event={"ID":"31fd23b6-fa18-4681-ba21-b9e69ad958d0","Type":"ContainerDied","Data":"92b02df8324395bd0df06a21434ae07095a74d3fba44daa130caad1fa833a873"} Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.484811 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" path="/var/lib/kubelet/pods/0a9f72e3-4a7d-4d44-a14f-f752a160ad3d/volumes" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.858811 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bvqd5"] Oct 09 08:34:34 crc kubenswrapper[4872]: E1009 08:34:34.859261 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="init" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.859292 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="init" Oct 09 08:34:34 crc kubenswrapper[4872]: E1009 08:34:34.859313 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e189a93b-9f2f-4c99-b315-69473291ed9c" containerName="mariadb-database-create" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.859324 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e189a93b-9f2f-4c99-b315-69473291ed9c" containerName="mariadb-database-create" Oct 09 08:34:34 crc kubenswrapper[4872]: E1009 08:34:34.859337 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="dnsmasq-dns" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.859349 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="dnsmasq-dns" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.859633 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e189a93b-9f2f-4c99-b315-69473291ed9c" containerName="mariadb-database-create" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.859700 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a9f72e3-4a7d-4d44-a14f-f752a160ad3d" containerName="dnsmasq-dns" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.860481 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.882816 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bvqd5"] Oct 09 08:34:34 crc kubenswrapper[4872]: I1009 08:34:34.959914 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62ps\" (UniqueName: \"kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps\") pod \"glance-db-create-bvqd5\" (UID: \"d67046a4-17d6-489f-857f-f8b016906baa\") " pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.061595 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f62ps\" (UniqueName: \"kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps\") pod \"glance-db-create-bvqd5\" (UID: \"d67046a4-17d6-489f-857f-f8b016906baa\") " pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.079043 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62ps\" (UniqueName: \"kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps\") pod \"glance-db-create-bvqd5\" (UID: \"d67046a4-17d6-489f-857f-f8b016906baa\") " pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.199410 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.646489 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bvqd5"] Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.754258 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887278 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887338 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887446 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887479 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpl59\" (UniqueName: \"kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887501 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887552 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.887580 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift\") pod \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\" (UID: \"31fd23b6-fa18-4681-ba21-b9e69ad958d0\") " Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.888690 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.889779 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.894507 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59" (OuterVolumeSpecName: "kube-api-access-fpl59") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "kube-api-access-fpl59". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.900753 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.909716 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts" (OuterVolumeSpecName: "scripts") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.916856 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.918751 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31fd23b6-fa18-4681-ba21-b9e69ad958d0" (UID: "31fd23b6-fa18-4681-ba21-b9e69ad958d0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991734 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991779 4872 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991793 4872 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/31fd23b6-fa18-4681-ba21-b9e69ad958d0-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991804 4872 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991817 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31fd23b6-fa18-4681-ba21-b9e69ad958d0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991828 4872 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/31fd23b6-fa18-4681-ba21-b9e69ad958d0-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:35 crc kubenswrapper[4872]: I1009 08:34:35.991838 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpl59\" (UniqueName: \"kubernetes.io/projected/31fd23b6-fa18-4681-ba21-b9e69ad958d0-kube-api-access-fpl59\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.006818 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.475343 4872 generic.go:334] "Generic (PLEG): container finished" podID="d67046a4-17d6-489f-857f-f8b016906baa" containerID="0b3af886e27dd94b6f2fcc82a81b968c48b231534811d8dfbb66ebf33c02e406" exitCode=0 Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.478243 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-qx7q8" Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.480206 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bvqd5" event={"ID":"d67046a4-17d6-489f-857f-f8b016906baa","Type":"ContainerDied","Data":"0b3af886e27dd94b6f2fcc82a81b968c48b231534811d8dfbb66ebf33c02e406"} Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.480250 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bvqd5" event={"ID":"d67046a4-17d6-489f-857f-f8b016906baa","Type":"ContainerStarted","Data":"0d34ea615ab58dd01a9f85f93d0058db85cb4ed96eb8ac53911b0d0393d033bd"} Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.480269 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-qx7q8" event={"ID":"31fd23b6-fa18-4681-ba21-b9e69ad958d0","Type":"ContainerDied","Data":"a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4"} Oct 09 08:34:36 crc kubenswrapper[4872]: I1009 08:34:36.480288 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6dde907948eebfa813ac8c5ce4d90b1a0ebe5e28cf052bbbd6fc55e185893e4" Oct 09 08:34:37 crc kubenswrapper[4872]: I1009 08:34:37.808187 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:37 crc kubenswrapper[4872]: I1009 08:34:37.921755 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f62ps\" (UniqueName: \"kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps\") pod \"d67046a4-17d6-489f-857f-f8b016906baa\" (UID: \"d67046a4-17d6-489f-857f-f8b016906baa\") " Oct 09 08:34:37 crc kubenswrapper[4872]: I1009 08:34:37.927782 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps" (OuterVolumeSpecName: "kube-api-access-f62ps") pod "d67046a4-17d6-489f-857f-f8b016906baa" (UID: "d67046a4-17d6-489f-857f-f8b016906baa"). InnerVolumeSpecName "kube-api-access-f62ps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.023764 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f62ps\" (UniqueName: \"kubernetes.io/projected/d67046a4-17d6-489f-857f-f8b016906baa-kube-api-access-f62ps\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.502466 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bvqd5" event={"ID":"d67046a4-17d6-489f-857f-f8b016906baa","Type":"ContainerDied","Data":"0d34ea615ab58dd01a9f85f93d0058db85cb4ed96eb8ac53911b0d0393d033bd"} Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.502521 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d34ea615ab58dd01a9f85f93d0058db85cb4ed96eb8ac53911b0d0393d033bd" Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.502531 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bvqd5" Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.734348 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:38 crc kubenswrapper[4872]: I1009 08:34:38.741804 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/6ee9bdf9-4380-47b9-b1b9-083974d8f01a-etc-swift\") pod \"swift-storage-0\" (UID: \"6ee9bdf9-4380-47b9-b1b9-083974d8f01a\") " pod="openstack/swift-storage-0" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.008701 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.326530 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-98xd2"] Oct 09 08:34:39 crc kubenswrapper[4872]: E1009 08:34:39.326992 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31fd23b6-fa18-4681-ba21-b9e69ad958d0" containerName="swift-ring-rebalance" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.327013 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="31fd23b6-fa18-4681-ba21-b9e69ad958d0" containerName="swift-ring-rebalance" Oct 09 08:34:39 crc kubenswrapper[4872]: E1009 08:34:39.327028 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67046a4-17d6-489f-857f-f8b016906baa" containerName="mariadb-database-create" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.327035 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67046a4-17d6-489f-857f-f8b016906baa" containerName="mariadb-database-create" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.327247 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="31fd23b6-fa18-4681-ba21-b9e69ad958d0" containerName="swift-ring-rebalance" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.327262 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67046a4-17d6-489f-857f-f8b016906baa" containerName="mariadb-database-create" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.327786 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.335193 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-98xd2"] Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.443838 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmfwn\" (UniqueName: \"kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn\") pod \"keystone-db-create-98xd2\" (UID: \"291de9b6-d184-4fdd-817a-ac679afcfe63\") " pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.545676 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmfwn\" (UniqueName: \"kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn\") pod \"keystone-db-create-98xd2\" (UID: \"291de9b6-d184-4fdd-817a-ac679afcfe63\") " pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.567359 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmfwn\" (UniqueName: \"kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn\") pod \"keystone-db-create-98xd2\" (UID: \"291de9b6-d184-4fdd-817a-ac679afcfe63\") " pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.597708 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 09 08:34:39 crc kubenswrapper[4872]: W1009 08:34:39.607768 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ee9bdf9_4380_47b9_b1b9_083974d8f01a.slice/crio-2bc1f17926a4c879b50379ac73bec3c813139541a56b23f0a7eba9205c400534 WatchSource:0}: Error finding container 2bc1f17926a4c879b50379ac73bec3c813139541a56b23f0a7eba9205c400534: Status 404 returned error can't find the container with id 2bc1f17926a4c879b50379ac73bec3c813139541a56b23f0a7eba9205c400534 Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.644873 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.749357 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-843f-account-create-kpbb9"] Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.752373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.755020 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.760939 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-843f-account-create-kpbb9"] Oct 09 08:34:39 crc kubenswrapper[4872]: I1009 08:34:39.951492 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6zk79\" (UniqueName: \"kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79\") pod \"placement-843f-account-create-kpbb9\" (UID: \"1ed1fcc1-b162-4818-abc1-4674bdf739c3\") " pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.054044 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6zk79\" (UniqueName: \"kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79\") pod \"placement-843f-account-create-kpbb9\" (UID: \"1ed1fcc1-b162-4818-abc1-4674bdf739c3\") " pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.087790 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6zk79\" (UniqueName: \"kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79\") pod \"placement-843f-account-create-kpbb9\" (UID: \"1ed1fcc1-b162-4818-abc1-4674bdf739c3\") " pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.104266 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-98xd2"] Oct 09 08:34:40 crc kubenswrapper[4872]: W1009 08:34:40.108159 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod291de9b6_d184_4fdd_817a_ac679afcfe63.slice/crio-ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020 WatchSource:0}: Error finding container ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020: Status 404 returned error can't find the container with id ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020 Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.380321 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.521137 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"2bc1f17926a4c879b50379ac73bec3c813139541a56b23f0a7eba9205c400534"} Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.523417 4872 generic.go:334] "Generic (PLEG): container finished" podID="291de9b6-d184-4fdd-817a-ac679afcfe63" containerID="556f0bc9baa9fe0435b1ccc929651c534bdbf7121e229eecc4eb288cc04af131" exitCode=0 Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.523484 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-98xd2" event={"ID":"291de9b6-d184-4fdd-817a-ac679afcfe63","Type":"ContainerDied","Data":"556f0bc9baa9fe0435b1ccc929651c534bdbf7121e229eecc4eb288cc04af131"} Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.523505 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-98xd2" event={"ID":"291de9b6-d184-4fdd-817a-ac679afcfe63","Type":"ContainerStarted","Data":"ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020"} Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.528977 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerDied","Data":"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c"} Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.530529 4872 generic.go:334] "Generic (PLEG): container finished" podID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerID="c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c" exitCode=0 Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.533125 4872 generic.go:334] "Generic (PLEG): container finished" podID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerID="4d374843dc49fd7757ddcf08ee23c47d6bb3f10cabe4e1deb7a461f2b2087d12" exitCode=0 Oct 09 08:34:40 crc kubenswrapper[4872]: I1009 08:34:40.533158 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerDied","Data":"4d374843dc49fd7757ddcf08ee23c47d6bb3f10cabe4e1deb7a461f2b2087d12"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.017335 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-843f-account-create-kpbb9"] Oct 09 08:34:41 crc kubenswrapper[4872]: W1009 08:34:41.040285 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ed1fcc1_b162_4818_abc1_4674bdf739c3.slice/crio-f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47 WatchSource:0}: Error finding container f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47: Status 404 returned error can't find the container with id f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47 Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.072340 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zk2nb" podUID="2e454e39-b253-4330-b5f7-7619c2f0f1c9" containerName="ovn-controller" probeResult="failure" output=< Oct 09 08:34:41 crc kubenswrapper[4872]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 09 08:34:41 crc kubenswrapper[4872]: > Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.544436 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"3c0d7ce1922a0a79aca541d0ab88e1b7300a890b7a1be27ae7b549011dbc3286"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.545020 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"f5a5a9e7094b1d5eeb77aee5f1733d1a13be8ab409aa7553c1457a8be4545649"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.545033 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"db24934d9c3dd0bababad43f1a9dc20456a13d6ed97ea43c2ca72a551b96fac8"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.546465 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerStarted","Data":"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.546772 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.548143 4872 generic.go:334] "Generic (PLEG): container finished" podID="1ed1fcc1-b162-4818-abc1-4674bdf739c3" containerID="33ee4fbb0d3c8e72f58c573536507b0c73b88f5a15ba3362f453b68cb6fc9a3b" exitCode=0 Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.548207 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-843f-account-create-kpbb9" event={"ID":"1ed1fcc1-b162-4818-abc1-4674bdf739c3","Type":"ContainerDied","Data":"33ee4fbb0d3c8e72f58c573536507b0c73b88f5a15ba3362f453b68cb6fc9a3b"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.548231 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-843f-account-create-kpbb9" event={"ID":"1ed1fcc1-b162-4818-abc1-4674bdf739c3","Type":"ContainerStarted","Data":"f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.556346 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerStarted","Data":"e8fa4656d598511564adf0e133ae4d8bb506b76e5cfa04100134020a2b209312"} Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.556831 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.587538 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=48.359322063 podStartE2EDuration="56.587517089s" podCreationTimestamp="2025-10-09 08:33:45 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.905509852 +0000 UTC m=+996.096038478" lastFinishedPulling="2025-10-09 08:34:06.133704868 +0000 UTC m=+1004.324233504" observedRunningTime="2025-10-09 08:34:41.578687954 +0000 UTC m=+1039.769216580" watchObservedRunningTime="2025-10-09 08:34:41.587517089 +0000 UTC m=+1039.778045715" Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.632686 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=49.051252063 podStartE2EDuration="57.632660742s" podCreationTimestamp="2025-10-09 08:33:44 +0000 UTC" firstStartedPulling="2025-10-09 08:33:57.604430939 +0000 UTC m=+995.794959565" lastFinishedPulling="2025-10-09 08:34:06.185839618 +0000 UTC m=+1004.376368244" observedRunningTime="2025-10-09 08:34:41.623398714 +0000 UTC m=+1039.813927360" watchObservedRunningTime="2025-10-09 08:34:41.632660742 +0000 UTC m=+1039.823189378" Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.857329 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.986938 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmfwn\" (UniqueName: \"kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn\") pod \"291de9b6-d184-4fdd-817a-ac679afcfe63\" (UID: \"291de9b6-d184-4fdd-817a-ac679afcfe63\") " Oct 09 08:34:41 crc kubenswrapper[4872]: I1009 08:34:41.994831 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn" (OuterVolumeSpecName: "kube-api-access-zmfwn") pod "291de9b6-d184-4fdd-817a-ac679afcfe63" (UID: "291de9b6-d184-4fdd-817a-ac679afcfe63"). InnerVolumeSpecName "kube-api-access-zmfwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.088301 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmfwn\" (UniqueName: \"kubernetes.io/projected/291de9b6-d184-4fdd-817a-ac679afcfe63-kube-api-access-zmfwn\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.574801 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"737941bdd7d0b87c3dccac2aa7eff0af7ba6a964641107d43849cdfd1eb838ed"} Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.578655 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-98xd2" event={"ID":"291de9b6-d184-4fdd-817a-ac679afcfe63","Type":"ContainerDied","Data":"ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020"} Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.578745 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-98xd2" Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.578796 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce5db195d8867e076ac402eb1debef5f8c4bf3866e5b0f8d5dd5aa42c03d2020" Oct 09 08:34:42 crc kubenswrapper[4872]: I1009 08:34:42.864349 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.006725 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6zk79\" (UniqueName: \"kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79\") pod \"1ed1fcc1-b162-4818-abc1-4674bdf739c3\" (UID: \"1ed1fcc1-b162-4818-abc1-4674bdf739c3\") " Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.011882 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79" (OuterVolumeSpecName: "kube-api-access-6zk79") pod "1ed1fcc1-b162-4818-abc1-4674bdf739c3" (UID: "1ed1fcc1-b162-4818-abc1-4674bdf739c3"). InnerVolumeSpecName "kube-api-access-6zk79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.108319 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6zk79\" (UniqueName: \"kubernetes.io/projected/1ed1fcc1-b162-4818-abc1-4674bdf739c3-kube-api-access-6zk79\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.588068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"743965717e0d476348998d95b0a4fe9cb566505f11255e6fdd083043e0ecf064"} Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.590163 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-843f-account-create-kpbb9" event={"ID":"1ed1fcc1-b162-4818-abc1-4674bdf739c3","Type":"ContainerDied","Data":"f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47"} Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.590205 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6b59468c7b3a420bf7ac5e0646566366be8d8adaf909d2447572d5a26f1ca47" Oct 09 08:34:43 crc kubenswrapper[4872]: I1009 08:34:43.590210 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-843f-account-create-kpbb9" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.601853 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"c3c6f0b360cd0eca2c8153bec2b3115ea24478c1e98e197a813a56cd2a0e248b"} Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.601910 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"4fe4d9181b2e86c397509b69651c2a01d44959ea18c35531b21d6832663a736e"} Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.956542 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-04fe-account-create-5nhvz"] Oct 09 08:34:44 crc kubenswrapper[4872]: E1009 08:34:44.957192 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ed1fcc1-b162-4818-abc1-4674bdf739c3" containerName="mariadb-account-create" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.957211 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ed1fcc1-b162-4818-abc1-4674bdf739c3" containerName="mariadb-account-create" Oct 09 08:34:44 crc kubenswrapper[4872]: E1009 08:34:44.957223 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="291de9b6-d184-4fdd-817a-ac679afcfe63" containerName="mariadb-database-create" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.957231 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="291de9b6-d184-4fdd-817a-ac679afcfe63" containerName="mariadb-database-create" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.957474 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ed1fcc1-b162-4818-abc1-4674bdf739c3" containerName="mariadb-account-create" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.957498 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="291de9b6-d184-4fdd-817a-ac679afcfe63" containerName="mariadb-database-create" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.958038 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.966871 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-04fe-account-create-5nhvz"] Oct 09 08:34:44 crc kubenswrapper[4872]: I1009 08:34:44.987852 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.043347 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzzkw\" (UniqueName: \"kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw\") pod \"glance-04fe-account-create-5nhvz\" (UID: \"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1\") " pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.144439 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzzkw\" (UniqueName: \"kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw\") pod \"glance-04fe-account-create-5nhvz\" (UID: \"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1\") " pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.167010 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzzkw\" (UniqueName: \"kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw\") pod \"glance-04fe-account-create-5nhvz\" (UID: \"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1\") " pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.297065 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.616004 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"a604b83443ea5fb39b14c65948f5a1fe31cba24edfc86c7d507d69ff393d6a71"} Oct 09 08:34:45 crc kubenswrapper[4872]: I1009 08:34:45.764671 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-04fe-account-create-5nhvz"] Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.072350 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zk2nb" podUID="2e454e39-b253-4330-b5f7-7619c2f0f1c9" containerName="ovn-controller" probeResult="failure" output=< Oct 09 08:34:46 crc kubenswrapper[4872]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 09 08:34:46 crc kubenswrapper[4872]: > Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.629970 4872 generic.go:334] "Generic (PLEG): container finished" podID="69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" containerID="3a64f434640ff8cb81b37264daa7527ee8f026aff3b786d914700dc7fbfe50bd" exitCode=0 Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.630072 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-04fe-account-create-5nhvz" event={"ID":"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1","Type":"ContainerDied","Data":"3a64f434640ff8cb81b37264daa7527ee8f026aff3b786d914700dc7fbfe50bd"} Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.630526 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-04fe-account-create-5nhvz" event={"ID":"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1","Type":"ContainerStarted","Data":"a44a4daab3f679f7ffbd03cceb971955910270e08ac593ad423e1bed47c9a733"} Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.643454 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"1fc0365449abd2f94f2cdcdbe62e89fda83e6f3da3061bdde306b13451d291ac"} Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.643506 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"a0cc6f561778372925aa41f9d30921bb75e4b0bd843c294a6d5ae21489fbf99c"} Oct 09 08:34:46 crc kubenswrapper[4872]: I1009 08:34:46.643522 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"a94075f04182c6c64abf2954befcc749dcdd8d709c52c16c59d7eff666c7a031"} Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.657756 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"bdc5e8c577a3a838c8a9748667f763cab8ec2db84533d56e607b76a226b8f62b"} Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.658062 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"66d5feb7148a58ab2c3bab18ac3510b10a300a13b23b6a26052af35e5665816f"} Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.658070 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"412dddf618ae117a8480d4a717968bbd97600c8a5d034004eb5b07bcf671e1e8"} Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.658078 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"6ee9bdf9-4380-47b9-b1b9-083974d8f01a","Type":"ContainerStarted","Data":"61138a161474037539a08b6a69534cbca6196fd0005d7b1cdaf9961ccb34d5ed"} Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.696227 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=20.299762405 podStartE2EDuration="26.696201589s" podCreationTimestamp="2025-10-09 08:34:21 +0000 UTC" firstStartedPulling="2025-10-09 08:34:39.61070354 +0000 UTC m=+1037.801232166" lastFinishedPulling="2025-10-09 08:34:46.007142724 +0000 UTC m=+1044.197671350" observedRunningTime="2025-10-09 08:34:47.686661963 +0000 UTC m=+1045.877190619" watchObservedRunningTime="2025-10-09 08:34:47.696201589 +0000 UTC m=+1045.886730215" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.956878 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.958409 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.962028 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.977137 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.993447 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.993823 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lszv\" (UniqueName: \"kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.993953 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.994100 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.994185 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:47 crc kubenswrapper[4872]: I1009 08:34:47.994226 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095684 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lszv\" (UniqueName: \"kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095778 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095825 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095866 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095895 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.095923 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.096791 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.096791 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.096957 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.097002 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.097028 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.116765 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lszv\" (UniqueName: \"kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv\") pod \"dnsmasq-dns-6d5b6d6b67-hlcbg\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.275058 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:48 crc kubenswrapper[4872]: I1009 08:34:48.948740 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.009539 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzzkw\" (UniqueName: \"kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw\") pod \"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1\" (UID: \"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1\") " Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.034217 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw" (OuterVolumeSpecName: "kube-api-access-tzzkw") pod "69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" (UID: "69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1"). InnerVolumeSpecName "kube-api-access-tzzkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.111513 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzzkw\" (UniqueName: \"kubernetes.io/projected/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1-kube-api-access-tzzkw\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.267242 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:34:49 crc kubenswrapper[4872]: W1009 08:34:49.273838 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod137de144_afdf_4553_be18_9a6e37243661.slice/crio-5b59872e86d6a7e5493b628eaa1aec4f4291d92cf65fe41a8c7d59a1513b546d WatchSource:0}: Error finding container 5b59872e86d6a7e5493b628eaa1aec4f4291d92cf65fe41a8c7d59a1513b546d: Status 404 returned error can't find the container with id 5b59872e86d6a7e5493b628eaa1aec4f4291d92cf65fe41a8c7d59a1513b546d Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.370793 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2695-account-create-rcqsk"] Oct 09 08:34:49 crc kubenswrapper[4872]: E1009 08:34:49.371402 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" containerName="mariadb-account-create" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.371478 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" containerName="mariadb-account-create" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.371745 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" containerName="mariadb-account-create" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.372450 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.374622 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.385496 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2695-account-create-rcqsk"] Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.430585 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbpg6\" (UniqueName: \"kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6\") pod \"keystone-2695-account-create-rcqsk\" (UID: \"0d4694ce-cbc2-4a3a-b46b-08049d1822d9\") " pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.532396 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbpg6\" (UniqueName: \"kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6\") pod \"keystone-2695-account-create-rcqsk\" (UID: \"0d4694ce-cbc2-4a3a-b46b-08049d1822d9\") " pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.550215 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbpg6\" (UniqueName: \"kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6\") pod \"keystone-2695-account-create-rcqsk\" (UID: \"0d4694ce-cbc2-4a3a-b46b-08049d1822d9\") " pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.671877 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-04fe-account-create-5nhvz" event={"ID":"69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1","Type":"ContainerDied","Data":"a44a4daab3f679f7ffbd03cceb971955910270e08ac593ad423e1bed47c9a733"} Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.671920 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a44a4daab3f679f7ffbd03cceb971955910270e08ac593ad423e1bed47c9a733" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.671898 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-04fe-account-create-5nhvz" Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.673298 4872 generic.go:334] "Generic (PLEG): container finished" podID="137de144-afdf-4553-be18-9a6e37243661" containerID="0cae0539d236193ce2d98f4eec1c06d55908d89ac56192dfcf07ad748919b750" exitCode=0 Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.673327 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" event={"ID":"137de144-afdf-4553-be18-9a6e37243661","Type":"ContainerDied","Data":"0cae0539d236193ce2d98f4eec1c06d55908d89ac56192dfcf07ad748919b750"} Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.673342 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" event={"ID":"137de144-afdf-4553-be18-9a6e37243661","Type":"ContainerStarted","Data":"5b59872e86d6a7e5493b628eaa1aec4f4291d92cf65fe41a8c7d59a1513b546d"} Oct 09 08:34:49 crc kubenswrapper[4872]: I1009 08:34:49.734799 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.158054 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2695-account-create-rcqsk"] Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.689851 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" event={"ID":"137de144-afdf-4553-be18-9a6e37243661","Type":"ContainerStarted","Data":"657b5aa5454781db9f6d0c43745909a179aa686d8bf45541b653a6696f7b3b42"} Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.690007 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.692055 4872 generic.go:334] "Generic (PLEG): container finished" podID="0d4694ce-cbc2-4a3a-b46b-08049d1822d9" containerID="8a0c592c04250cdba12ebb992b64f712f9b5e1aefe15e21c76ea554cac59e51e" exitCode=0 Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.692148 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2695-account-create-rcqsk" event={"ID":"0d4694ce-cbc2-4a3a-b46b-08049d1822d9","Type":"ContainerDied","Data":"8a0c592c04250cdba12ebb992b64f712f9b5e1aefe15e21c76ea554cac59e51e"} Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.692398 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2695-account-create-rcqsk" event={"ID":"0d4694ce-cbc2-4a3a-b46b-08049d1822d9","Type":"ContainerStarted","Data":"0a80e7e94f0ff1d7b6b1516618a810aaf2b6757aab2b1213152d2d35ca4dbc50"} Oct 09 08:34:50 crc kubenswrapper[4872]: I1009 08:34:50.714995 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" podStartSLOduration=3.714968137 podStartE2EDuration="3.714968137s" podCreationTimestamp="2025-10-09 08:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:34:50.709085747 +0000 UTC m=+1048.899614383" watchObservedRunningTime="2025-10-09 08:34:50.714968137 +0000 UTC m=+1048.905496773" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.068278 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-zk2nb" podUID="2e454e39-b253-4330-b5f7-7619c2f0f1c9" containerName="ovn-controller" probeResult="failure" output=< Oct 09 08:34:51 crc kubenswrapper[4872]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 09 08:34:51 crc kubenswrapper[4872]: > Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.081845 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.101120 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mfxrr" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.315813 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-zk2nb-config-cjnk5"] Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.316832 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.318512 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.330166 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zk2nb-config-cjnk5"] Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.377970 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.378063 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.378110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.378184 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.378209 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stbz4\" (UniqueName: \"kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.378279 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.479972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480105 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480163 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480228 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stbz4\" (UniqueName: \"kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480347 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480372 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.480362 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.481584 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.482938 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.502170 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stbz4\" (UniqueName: \"kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4\") pod \"ovn-controller-zk2nb-config-cjnk5\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:51 crc kubenswrapper[4872]: I1009 08:34:51.646813 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.047335 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.193719 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbpg6\" (UniqueName: \"kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6\") pod \"0d4694ce-cbc2-4a3a-b46b-08049d1822d9\" (UID: \"0d4694ce-cbc2-4a3a-b46b-08049d1822d9\") " Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.195007 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-zk2nb-config-cjnk5"] Oct 09 08:34:52 crc kubenswrapper[4872]: W1009 08:34:52.199374 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod097e3ca9_8c24_40ef_9c82_66b901aa4e9c.slice/crio-99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1 WatchSource:0}: Error finding container 99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1: Status 404 returned error can't find the container with id 99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1 Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.199653 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6" (OuterVolumeSpecName: "kube-api-access-gbpg6") pod "0d4694ce-cbc2-4a3a-b46b-08049d1822d9" (UID: "0d4694ce-cbc2-4a3a-b46b-08049d1822d9"). InnerVolumeSpecName "kube-api-access-gbpg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.295928 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbpg6\" (UniqueName: \"kubernetes.io/projected/0d4694ce-cbc2-4a3a-b46b-08049d1822d9-kube-api-access-gbpg6\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.716989 4872 generic.go:334] "Generic (PLEG): container finished" podID="097e3ca9-8c24-40ef-9c82-66b901aa4e9c" containerID="5c83c3b24f32f5e798e35bf466c8d371ec7a7489108dacfe9944717d98ef6cde" exitCode=0 Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.717054 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zk2nb-config-cjnk5" event={"ID":"097e3ca9-8c24-40ef-9c82-66b901aa4e9c","Type":"ContainerDied","Data":"5c83c3b24f32f5e798e35bf466c8d371ec7a7489108dacfe9944717d98ef6cde"} Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.717261 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zk2nb-config-cjnk5" event={"ID":"097e3ca9-8c24-40ef-9c82-66b901aa4e9c","Type":"ContainerStarted","Data":"99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1"} Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.719611 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2695-account-create-rcqsk" event={"ID":"0d4694ce-cbc2-4a3a-b46b-08049d1822d9","Type":"ContainerDied","Data":"0a80e7e94f0ff1d7b6b1516618a810aaf2b6757aab2b1213152d2d35ca4dbc50"} Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.719668 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a80e7e94f0ff1d7b6b1516618a810aaf2b6757aab2b1213152d2d35ca4dbc50" Oct 09 08:34:52 crc kubenswrapper[4872]: I1009 08:34:52.719683 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2695-account-create-rcqsk" Oct 09 08:34:53 crc kubenswrapper[4872]: I1009 08:34:53.998259 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126294 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126361 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126489 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126520 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stbz4\" (UniqueName: \"kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts\") pod \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\" (UID: \"097e3ca9-8c24-40ef-9c82-66b901aa4e9c\") " Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.126621 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run" (OuterVolumeSpecName: "var-run") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127234 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127454 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts" (OuterVolumeSpecName: "scripts") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127800 4872 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127819 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127827 4872 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127835 4872 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.127846 4872 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.132369 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4" (OuterVolumeSpecName: "kube-api-access-stbz4") pod "097e3ca9-8c24-40ef-9c82-66b901aa4e9c" (UID: "097e3ca9-8c24-40ef-9c82-66b901aa4e9c"). InnerVolumeSpecName "kube-api-access-stbz4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.229768 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stbz4\" (UniqueName: \"kubernetes.io/projected/097e3ca9-8c24-40ef-9c82-66b901aa4e9c-kube-api-access-stbz4\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.735666 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-zk2nb-config-cjnk5" event={"ID":"097e3ca9-8c24-40ef-9c82-66b901aa4e9c","Type":"ContainerDied","Data":"99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1"} Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.736013 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99879b06e06ee547932192e5e4bca1a5bd96a94a3204904892a92ae5c6a0a4a1" Oct 09 08:34:54 crc kubenswrapper[4872]: I1009 08:34:54.735696 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-zk2nb-config-cjnk5" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.105192 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-zk2nb-config-cjnk5"] Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.113160 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-zk2nb-config-cjnk5"] Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.123490 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-jvtzz"] Oct 09 08:34:55 crc kubenswrapper[4872]: E1009 08:34:55.123907 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d4694ce-cbc2-4a3a-b46b-08049d1822d9" containerName="mariadb-account-create" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.123924 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d4694ce-cbc2-4a3a-b46b-08049d1822d9" containerName="mariadb-account-create" Oct 09 08:34:55 crc kubenswrapper[4872]: E1009 08:34:55.123955 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097e3ca9-8c24-40ef-9c82-66b901aa4e9c" containerName="ovn-config" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.123962 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="097e3ca9-8c24-40ef-9c82-66b901aa4e9c" containerName="ovn-config" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.124135 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d4694ce-cbc2-4a3a-b46b-08049d1822d9" containerName="mariadb-account-create" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.124172 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="097e3ca9-8c24-40ef-9c82-66b901aa4e9c" containerName="ovn-config" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.128820 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.131110 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.132021 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-sk5xf" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.142818 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jvtzz"] Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.251320 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2chm\" (UniqueName: \"kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.251401 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.251442 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.251479 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.352864 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2chm\" (UniqueName: \"kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.352942 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.352976 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.353017 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.357909 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.358343 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.358949 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.376220 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2chm\" (UniqueName: \"kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm\") pod \"glance-db-sync-jvtzz\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.450712 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvtzz" Oct 09 08:34:55 crc kubenswrapper[4872]: I1009 08:34:55.971609 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-jvtzz"] Oct 09 08:34:55 crc kubenswrapper[4872]: W1009 08:34:55.979038 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod256fd449_62bd_44fe_945e_92663ab62698.slice/crio-2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391 WatchSource:0}: Error finding container 2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391: Status 404 returned error can't find the container with id 2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391 Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.071229 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-zk2nb" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.290337 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.473997 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097e3ca9-8c24-40ef-9c82-66b901aa4e9c" path="/var/lib/kubelet/pods/097e3ca9-8c24-40ef-9c82-66b901aa4e9c/volumes" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.599872 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-8xw9z"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.601016 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8xw9z" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.617487 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8xw9z"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.676912 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48dqp\" (UniqueName: \"kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp\") pod \"cinder-db-create-8xw9z\" (UID: \"3231028e-ac1b-4f07-9bc6-3efb4160cfb0\") " pod="openstack/cinder-db-create-8xw9z" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.693854 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-4x54s"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.695000 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.708006 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4x54s"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.750268 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.753946 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvtzz" event={"ID":"256fd449-62bd-44fe-945e-92663ab62698","Type":"ContainerStarted","Data":"2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391"} Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.778265 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48dqp\" (UniqueName: \"kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp\") pod \"cinder-db-create-8xw9z\" (UID: \"3231028e-ac1b-4f07-9bc6-3efb4160cfb0\") " pod="openstack/cinder-db-create-8xw9z" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.778354 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sl5kz\" (UniqueName: \"kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz\") pod \"barbican-db-create-4x54s\" (UID: \"fca98354-3506-4c2a-99d3-f482190c4326\") " pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.816443 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48dqp\" (UniqueName: \"kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp\") pod \"cinder-db-create-8xw9z\" (UID: \"3231028e-ac1b-4f07-9bc6-3efb4160cfb0\") " pod="openstack/cinder-db-create-8xw9z" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.880621 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-wwgjr"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.880785 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sl5kz\" (UniqueName: \"kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz\") pod \"barbican-db-create-4x54s\" (UID: \"fca98354-3506-4c2a-99d3-f482190c4326\") " pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.882775 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.885715 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.892124 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kvgxv" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.892152 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.892508 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.897181 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wwgjr"] Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.914340 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sl5kz\" (UniqueName: \"kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz\") pod \"barbican-db-create-4x54s\" (UID: \"fca98354-3506-4c2a-99d3-f482190c4326\") " pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.928439 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8xw9z" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.982996 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.983090 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7lgf\" (UniqueName: \"kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:56 crc kubenswrapper[4872]: I1009 08:34:56.983129 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.001253 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-cp4rx"] Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.002363 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cp4rx" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.016353 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.027462 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cp4rx"] Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.084034 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.084081 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdr9b\" (UniqueName: \"kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b\") pod \"neutron-db-create-cp4rx\" (UID: \"7407ecf1-4c1b-47cd-9c8c-099630211671\") " pod="openstack/neutron-db-create-cp4rx" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.084141 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7lgf\" (UniqueName: \"kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.084173 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.087465 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.087913 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.104708 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7lgf\" (UniqueName: \"kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf\") pod \"keystone-db-sync-wwgjr\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.186056 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdr9b\" (UniqueName: \"kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b\") pod \"neutron-db-create-cp4rx\" (UID: \"7407ecf1-4c1b-47cd-9c8c-099630211671\") " pod="openstack/neutron-db-create-cp4rx" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.205538 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.209861 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdr9b\" (UniqueName: \"kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b\") pod \"neutron-db-create-cp4rx\" (UID: \"7407ecf1-4c1b-47cd-9c8c-099630211671\") " pod="openstack/neutron-db-create-cp4rx" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.277766 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-4x54s"] Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.387824 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cp4rx" Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.526924 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-8xw9z"] Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.661805 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-wwgjr"] Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.762370 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wwgjr" event={"ID":"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c","Type":"ContainerStarted","Data":"a7d87d3642884141e5b7fbc72064aea61002a4b0453d8c90cd94395b8ca5c824"} Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.763609 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8xw9z" event={"ID":"3231028e-ac1b-4f07-9bc6-3efb4160cfb0","Type":"ContainerStarted","Data":"1fb3f157b4a8b1396a8744b63e9091178924b6bdd70131394bf7365e425f3739"} Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.764881 4872 generic.go:334] "Generic (PLEG): container finished" podID="fca98354-3506-4c2a-99d3-f482190c4326" containerID="ce0c4c5c5b53ea1f12864087e6d9a473e854d32244f134fe3f0db4ca61e58a02" exitCode=0 Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.764915 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4x54s" event={"ID":"fca98354-3506-4c2a-99d3-f482190c4326","Type":"ContainerDied","Data":"ce0c4c5c5b53ea1f12864087e6d9a473e854d32244f134fe3f0db4ca61e58a02"} Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.764932 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4x54s" event={"ID":"fca98354-3506-4c2a-99d3-f482190c4326","Type":"ContainerStarted","Data":"e5b372d294c60e28bcd1d30a5edfb4eda205826d9c95e3ab6b4606ebfcf71b70"} Oct 09 08:34:57 crc kubenswrapper[4872]: I1009 08:34:57.881948 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-cp4rx"] Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.276776 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.324528 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.324766 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="dnsmasq-dns" containerID="cri-o://1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724" gracePeriod=10 Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.756694 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.777526 4872 generic.go:334] "Generic (PLEG): container finished" podID="7407ecf1-4c1b-47cd-9c8c-099630211671" containerID="c68ade29cc29c658f9e26ea63dce63dea34e96435930126fdaf774d09d87827b" exitCode=0 Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.777604 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cp4rx" event={"ID":"7407ecf1-4c1b-47cd-9c8c-099630211671","Type":"ContainerDied","Data":"c68ade29cc29c658f9e26ea63dce63dea34e96435930126fdaf774d09d87827b"} Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.777630 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cp4rx" event={"ID":"7407ecf1-4c1b-47cd-9c8c-099630211671","Type":"ContainerStarted","Data":"0317d4c5e61a079c70735ca33d39959f5735b25dc0a9730475dce9f0e9bf9c67"} Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.791906 4872 generic.go:334] "Generic (PLEG): container finished" podID="3231028e-ac1b-4f07-9bc6-3efb4160cfb0" containerID="12437e4f66176fed9155f6ea28e39b7f9b4efd6613206c6839dfa04e52d476b3" exitCode=0 Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.791974 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8xw9z" event={"ID":"3231028e-ac1b-4f07-9bc6-3efb4160cfb0","Type":"ContainerDied","Data":"12437e4f66176fed9155f6ea28e39b7f9b4efd6613206c6839dfa04e52d476b3"} Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.801863 4872 generic.go:334] "Generic (PLEG): container finished" podID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerID="1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724" exitCode=0 Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.802010 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.802074 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" event={"ID":"ad0e49d0-0cc0-4f1d-8212-f798c25d4017","Type":"ContainerDied","Data":"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724"} Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.802107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-gl5sb" event={"ID":"ad0e49d0-0cc0-4f1d-8212-f798c25d4017","Type":"ContainerDied","Data":"0e71c37dd14d299f3a264f4b109506c349658e9a5cec5ee261a4155da0f4a037"} Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.802127 4872 scope.go:117] "RemoveContainer" containerID="1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.814661 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb\") pod \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.814857 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc\") pod \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.814899 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb\") pod \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.814946 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config\") pod \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.814993 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5h66p\" (UniqueName: \"kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p\") pod \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\" (UID: \"ad0e49d0-0cc0-4f1d-8212-f798c25d4017\") " Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.823177 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p" (OuterVolumeSpecName: "kube-api-access-5h66p") pod "ad0e49d0-0cc0-4f1d-8212-f798c25d4017" (UID: "ad0e49d0-0cc0-4f1d-8212-f798c25d4017"). InnerVolumeSpecName "kube-api-access-5h66p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.870797 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ad0e49d0-0cc0-4f1d-8212-f798c25d4017" (UID: "ad0e49d0-0cc0-4f1d-8212-f798c25d4017"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.886944 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ad0e49d0-0cc0-4f1d-8212-f798c25d4017" (UID: "ad0e49d0-0cc0-4f1d-8212-f798c25d4017"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.897130 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ad0e49d0-0cc0-4f1d-8212-f798c25d4017" (UID: "ad0e49d0-0cc0-4f1d-8212-f798c25d4017"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.899786 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config" (OuterVolumeSpecName: "config") pod "ad0e49d0-0cc0-4f1d-8212-f798c25d4017" (UID: "ad0e49d0-0cc0-4f1d-8212-f798c25d4017"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.917262 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.917302 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.917316 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.917327 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5h66p\" (UniqueName: \"kubernetes.io/projected/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-kube-api-access-5h66p\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.917339 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ad0e49d0-0cc0-4f1d-8212-f798c25d4017-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:58 crc kubenswrapper[4872]: I1009 08:34:58.967451 4872 scope.go:117] "RemoveContainer" containerID="3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.016148 4872 scope.go:117] "RemoveContainer" containerID="1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724" Oct 09 08:34:59 crc kubenswrapper[4872]: E1009 08:34:59.016622 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724\": container with ID starting with 1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724 not found: ID does not exist" containerID="1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.016687 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724"} err="failed to get container status \"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724\": rpc error: code = NotFound desc = could not find container \"1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724\": container with ID starting with 1e2242bc2fd95b0457c80348d75633c5016e4fb65f49ed86b7806242df4fa724 not found: ID does not exist" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.016708 4872 scope.go:117] "RemoveContainer" containerID="3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad" Oct 09 08:34:59 crc kubenswrapper[4872]: E1009 08:34:59.025272 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad\": container with ID starting with 3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad not found: ID does not exist" containerID="3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.025318 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad"} err="failed to get container status \"3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad\": rpc error: code = NotFound desc = could not find container \"3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad\": container with ID starting with 3b1789d4c00a65344870f34d3367fb83090dbcae859bf281b5710dd1b487a4ad not found: ID does not exist" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.156958 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.179694 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.189531 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-gl5sb"] Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.221828 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sl5kz\" (UniqueName: \"kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz\") pod \"fca98354-3506-4c2a-99d3-f482190c4326\" (UID: \"fca98354-3506-4c2a-99d3-f482190c4326\") " Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.228425 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz" (OuterVolumeSpecName: "kube-api-access-sl5kz") pod "fca98354-3506-4c2a-99d3-f482190c4326" (UID: "fca98354-3506-4c2a-99d3-f482190c4326"). InnerVolumeSpecName "kube-api-access-sl5kz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.323582 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sl5kz\" (UniqueName: \"kubernetes.io/projected/fca98354-3506-4c2a-99d3-f482190c4326-kube-api-access-sl5kz\") on node \"crc\" DevicePath \"\"" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.812529 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-4x54s" Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.812498 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-4x54s" event={"ID":"fca98354-3506-4c2a-99d3-f482190c4326","Type":"ContainerDied","Data":"e5b372d294c60e28bcd1d30a5edfb4eda205826d9c95e3ab6b4606ebfcf71b70"} Oct 09 08:34:59 crc kubenswrapper[4872]: I1009 08:34:59.813512 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5b372d294c60e28bcd1d30a5edfb4eda205826d9c95e3ab6b4606ebfcf71b70" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.184272 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cp4rx" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.190234 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8xw9z" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.251507 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdr9b\" (UniqueName: \"kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b\") pod \"7407ecf1-4c1b-47cd-9c8c-099630211671\" (UID: \"7407ecf1-4c1b-47cd-9c8c-099630211671\") " Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.251736 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48dqp\" (UniqueName: \"kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp\") pod \"3231028e-ac1b-4f07-9bc6-3efb4160cfb0\" (UID: \"3231028e-ac1b-4f07-9bc6-3efb4160cfb0\") " Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.258023 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b" (OuterVolumeSpecName: "kube-api-access-fdr9b") pod "7407ecf1-4c1b-47cd-9c8c-099630211671" (UID: "7407ecf1-4c1b-47cd-9c8c-099630211671"). InnerVolumeSpecName "kube-api-access-fdr9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.258788 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp" (OuterVolumeSpecName: "kube-api-access-48dqp") pod "3231028e-ac1b-4f07-9bc6-3efb4160cfb0" (UID: "3231028e-ac1b-4f07-9bc6-3efb4160cfb0"). InnerVolumeSpecName "kube-api-access-48dqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.353971 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48dqp\" (UniqueName: \"kubernetes.io/projected/3231028e-ac1b-4f07-9bc6-3efb4160cfb0-kube-api-access-48dqp\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.354003 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdr9b\" (UniqueName: \"kubernetes.io/projected/7407ecf1-4c1b-47cd-9c8c-099630211671-kube-api-access-fdr9b\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.475249 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" path="/var/lib/kubelet/pods/ad0e49d0-0cc0-4f1d-8212-f798c25d4017/volumes" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.826816 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-8xw9z" event={"ID":"3231028e-ac1b-4f07-9bc6-3efb4160cfb0","Type":"ContainerDied","Data":"1fb3f157b4a8b1396a8744b63e9091178924b6bdd70131394bf7365e425f3739"} Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.826868 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1fb3f157b4a8b1396a8744b63e9091178924b6bdd70131394bf7365e425f3739" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.826827 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-8xw9z" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.828836 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-cp4rx" Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.828756 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-cp4rx" event={"ID":"7407ecf1-4c1b-47cd-9c8c-099630211671","Type":"ContainerDied","Data":"0317d4c5e61a079c70735ca33d39959f5735b25dc0a9730475dce9f0e9bf9c67"} Oct 09 08:35:00 crc kubenswrapper[4872]: I1009 08:35:00.829029 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0317d4c5e61a079c70735ca33d39959f5735b25dc0a9730475dce9f0e9bf9c67" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.751279 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-cef4-account-create-nr247"] Oct 09 08:35:06 crc kubenswrapper[4872]: E1009 08:35:06.752249 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3231028e-ac1b-4f07-9bc6-3efb4160cfb0" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752266 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3231028e-ac1b-4f07-9bc6-3efb4160cfb0" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: E1009 08:35:06.752278 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca98354-3506-4c2a-99d3-f482190c4326" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752286 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca98354-3506-4c2a-99d3-f482190c4326" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: E1009 08:35:06.752308 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7407ecf1-4c1b-47cd-9c8c-099630211671" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752315 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7407ecf1-4c1b-47cd-9c8c-099630211671" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: E1009 08:35:06.752331 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="dnsmasq-dns" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752338 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="dnsmasq-dns" Oct 09 08:35:06 crc kubenswrapper[4872]: E1009 08:35:06.752350 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="init" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752358 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="init" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752557 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca98354-3506-4c2a-99d3-f482190c4326" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752579 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad0e49d0-0cc0-4f1d-8212-f798c25d4017" containerName="dnsmasq-dns" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752604 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3231028e-ac1b-4f07-9bc6-3efb4160cfb0" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.752615 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="7407ecf1-4c1b-47cd-9c8c-099630211671" containerName="mariadb-database-create" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.753517 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.757582 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.761022 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cef4-account-create-nr247"] Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.880301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z48vq\" (UniqueName: \"kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq\") pod \"cinder-cef4-account-create-nr247\" (UID: \"b3032b0c-ddae-450b-be63-1cd4b9014afb\") " pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.933895 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-ce3d-account-create-65snb"] Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.936691 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.939444 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.944144 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ce3d-account-create-65snb"] Oct 09 08:35:06 crc kubenswrapper[4872]: I1009 08:35:06.982310 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z48vq\" (UniqueName: \"kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq\") pod \"cinder-cef4-account-create-nr247\" (UID: \"b3032b0c-ddae-450b-be63-1cd4b9014afb\") " pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.000632 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z48vq\" (UniqueName: \"kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq\") pod \"cinder-cef4-account-create-nr247\" (UID: \"b3032b0c-ddae-450b-be63-1cd4b9014afb\") " pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.040012 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-8678-account-create-xs9w6"] Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.041177 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.043302 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.049259 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8678-account-create-xs9w6"] Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.077033 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.083396 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2zzj\" (UniqueName: \"kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj\") pod \"barbican-ce3d-account-create-65snb\" (UID: \"b6154eba-fb06-4583-8937-9c1183083fec\") " pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.186711 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2zzj\" (UniqueName: \"kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj\") pod \"barbican-ce3d-account-create-65snb\" (UID: \"b6154eba-fb06-4583-8937-9c1183083fec\") " pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.186865 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5qzx\" (UniqueName: \"kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx\") pod \"neutron-8678-account-create-xs9w6\" (UID: \"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d\") " pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.209421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2zzj\" (UniqueName: \"kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj\") pod \"barbican-ce3d-account-create-65snb\" (UID: \"b6154eba-fb06-4583-8937-9c1183083fec\") " pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.257679 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.288685 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5qzx\" (UniqueName: \"kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx\") pod \"neutron-8678-account-create-xs9w6\" (UID: \"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d\") " pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.307331 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5qzx\" (UniqueName: \"kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx\") pod \"neutron-8678-account-create-xs9w6\" (UID: \"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d\") " pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:07 crc kubenswrapper[4872]: I1009 08:35:07.368076 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:09 crc kubenswrapper[4872]: I1009 08:35:09.899310 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wwgjr" event={"ID":"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c","Type":"ContainerStarted","Data":"045016146c016ce9c181dfc7b589775fb105549bf098f7943aeda130b5eaf38f"} Oct 09 08:35:09 crc kubenswrapper[4872]: I1009 08:35:09.919088 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-wwgjr" podStartSLOduration=2.033340618 podStartE2EDuration="13.919071488s" podCreationTimestamp="2025-10-09 08:34:56 +0000 UTC" firstStartedPulling="2025-10-09 08:34:57.689795453 +0000 UTC m=+1055.880324079" lastFinishedPulling="2025-10-09 08:35:09.575526323 +0000 UTC m=+1067.766054949" observedRunningTime="2025-10-09 08:35:09.912691434 +0000 UTC m=+1068.103220050" watchObservedRunningTime="2025-10-09 08:35:09.919071488 +0000 UTC m=+1068.109600114" Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.106338 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-8678-account-create-xs9w6"] Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.113977 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-ce3d-account-create-65snb"] Oct 09 08:35:10 crc kubenswrapper[4872]: W1009 08:35:10.118212 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c611709_a037_41b7_a8f6_1dc9fdbb8d4d.slice/crio-2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00 WatchSource:0}: Error finding container 2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00: Status 404 returned error can't find the container with id 2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00 Oct 09 08:35:10 crc kubenswrapper[4872]: W1009 08:35:10.120332 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6154eba_fb06_4583_8937_9c1183083fec.slice/crio-fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971 WatchSource:0}: Error finding container fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971: Status 404 returned error can't find the container with id fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971 Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.229009 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-cef4-account-create-nr247"] Oct 09 08:35:10 crc kubenswrapper[4872]: W1009 08:35:10.234414 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb3032b0c_ddae_450b_be63_1cd4b9014afb.slice/crio-40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c WatchSource:0}: Error finding container 40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c: Status 404 returned error can't find the container with id 40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.911397 4872 generic.go:334] "Generic (PLEG): container finished" podID="1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" containerID="5085d9b21dfba4b5228d6fa27fd7273de62f29732f8645437c5109b0d928d618" exitCode=0 Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.911759 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8678-account-create-xs9w6" event={"ID":"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d","Type":"ContainerDied","Data":"5085d9b21dfba4b5228d6fa27fd7273de62f29732f8645437c5109b0d928d618"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.911816 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8678-account-create-xs9w6" event={"ID":"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d","Type":"ContainerStarted","Data":"2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.913389 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvtzz" event={"ID":"256fd449-62bd-44fe-945e-92663ab62698","Type":"ContainerStarted","Data":"fc5356d5eb0b02f29e7d0e03b74c7bf865150c941377f1311308f3be7e10c087"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.917502 4872 generic.go:334] "Generic (PLEG): container finished" podID="b6154eba-fb06-4583-8937-9c1183083fec" containerID="84e6959f826be17cb56ea1b113ca20739fa8c18b27e7ee8b3e9e2559c7d093b8" exitCode=0 Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.917577 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce3d-account-create-65snb" event={"ID":"b6154eba-fb06-4583-8937-9c1183083fec","Type":"ContainerDied","Data":"84e6959f826be17cb56ea1b113ca20739fa8c18b27e7ee8b3e9e2559c7d093b8"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.917609 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce3d-account-create-65snb" event={"ID":"b6154eba-fb06-4583-8937-9c1183083fec","Type":"ContainerStarted","Data":"fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.919272 4872 generic.go:334] "Generic (PLEG): container finished" podID="b3032b0c-ddae-450b-be63-1cd4b9014afb" containerID="f4b32ad768096201576d9394662a22fb88db28900c4212fa60dd91a498d8bc6b" exitCode=0 Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.921583 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cef4-account-create-nr247" event={"ID":"b3032b0c-ddae-450b-be63-1cd4b9014afb","Type":"ContainerDied","Data":"f4b32ad768096201576d9394662a22fb88db28900c4212fa60dd91a498d8bc6b"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.921681 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cef4-account-create-nr247" event={"ID":"b3032b0c-ddae-450b-be63-1cd4b9014afb","Type":"ContainerStarted","Data":"40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c"} Oct 09 08:35:10 crc kubenswrapper[4872]: I1009 08:35:10.975545 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-jvtzz" podStartSLOduration=2.33621648 podStartE2EDuration="15.975525745s" podCreationTimestamp="2025-10-09 08:34:55 +0000 UTC" firstStartedPulling="2025-10-09 08:34:55.98203242 +0000 UTC m=+1054.172561056" lastFinishedPulling="2025-10-09 08:35:09.621341695 +0000 UTC m=+1067.811870321" observedRunningTime="2025-10-09 08:35:10.973726763 +0000 UTC m=+1069.164255389" watchObservedRunningTime="2025-10-09 08:35:10.975525745 +0000 UTC m=+1069.166054381" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.408023 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.489734 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2zzj\" (UniqueName: \"kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj\") pod \"b6154eba-fb06-4583-8937-9c1183083fec\" (UID: \"b6154eba-fb06-4583-8937-9c1183083fec\") " Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.495861 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj" (OuterVolumeSpecName: "kube-api-access-l2zzj") pod "b6154eba-fb06-4583-8937-9c1183083fec" (UID: "b6154eba-fb06-4583-8937-9c1183083fec"). InnerVolumeSpecName "kube-api-access-l2zzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.542221 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.548587 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.591374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5qzx\" (UniqueName: \"kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx\") pod \"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d\" (UID: \"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d\") " Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.591703 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z48vq\" (UniqueName: \"kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq\") pod \"b3032b0c-ddae-450b-be63-1cd4b9014afb\" (UID: \"b3032b0c-ddae-450b-be63-1cd4b9014afb\") " Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.592221 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2zzj\" (UniqueName: \"kubernetes.io/projected/b6154eba-fb06-4583-8937-9c1183083fec-kube-api-access-l2zzj\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.594240 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx" (OuterVolumeSpecName: "kube-api-access-d5qzx") pod "1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" (UID: "1c611709-a037-41b7-a8f6-1dc9fdbb8d4d"). InnerVolumeSpecName "kube-api-access-d5qzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.594861 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq" (OuterVolumeSpecName: "kube-api-access-z48vq") pod "b3032b0c-ddae-450b-be63-1cd4b9014afb" (UID: "b3032b0c-ddae-450b-be63-1cd4b9014afb"). InnerVolumeSpecName "kube-api-access-z48vq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.694325 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5qzx\" (UniqueName: \"kubernetes.io/projected/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d-kube-api-access-d5qzx\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.694363 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z48vq\" (UniqueName: \"kubernetes.io/projected/b3032b0c-ddae-450b-be63-1cd4b9014afb-kube-api-access-z48vq\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.934224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-8678-account-create-xs9w6" event={"ID":"1c611709-a037-41b7-a8f6-1dc9fdbb8d4d","Type":"ContainerDied","Data":"2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00"} Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.934269 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2e04a8c9e95136a5f8e2db15b74713c2949d7f7018ea9a215438344a54ff8e00" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.934236 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-8678-account-create-xs9w6" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.936709 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-ce3d-account-create-65snb" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.936712 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-ce3d-account-create-65snb" event={"ID":"b6154eba-fb06-4583-8937-9c1183083fec","Type":"ContainerDied","Data":"fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971"} Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.936848 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa44fd826d44ddd307c870d956ea74431d3054f517d490787e4876da0e86a971" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.938394 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-cef4-account-create-nr247" event={"ID":"b3032b0c-ddae-450b-be63-1cd4b9014afb","Type":"ContainerDied","Data":"40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c"} Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.938424 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40e2d2c8e86048016e18b0c60cdfb7ffe3287d0591aa280d924e1d2696b9853c" Oct 09 08:35:12 crc kubenswrapper[4872]: I1009 08:35:12.938437 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-cef4-account-create-nr247" Oct 09 08:35:13 crc kubenswrapper[4872]: I1009 08:35:13.950440 4872 generic.go:334] "Generic (PLEG): container finished" podID="f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" containerID="045016146c016ce9c181dfc7b589775fb105549bf098f7943aeda130b5eaf38f" exitCode=0 Oct 09 08:35:13 crc kubenswrapper[4872]: I1009 08:35:13.950484 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wwgjr" event={"ID":"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c","Type":"ContainerDied","Data":"045016146c016ce9c181dfc7b589775fb105549bf098f7943aeda130b5eaf38f"} Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.269680 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.335984 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7lgf\" (UniqueName: \"kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf\") pod \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.336061 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle\") pod \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.336100 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data\") pod \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\" (UID: \"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c\") " Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.341977 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf" (OuterVolumeSpecName: "kube-api-access-x7lgf") pod "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" (UID: "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c"). InnerVolumeSpecName "kube-api-access-x7lgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.368503 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" (UID: "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.393152 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data" (OuterVolumeSpecName: "config-data") pod "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" (UID: "f5134c98-fe5f-4fba-ac37-e3c32b0ba26c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.437776 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7lgf\" (UniqueName: \"kubernetes.io/projected/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-kube-api-access-x7lgf\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.437848 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.437858 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.971966 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-wwgjr" event={"ID":"f5134c98-fe5f-4fba-ac37-e3c32b0ba26c","Type":"ContainerDied","Data":"a7d87d3642884141e5b7fbc72064aea61002a4b0453d8c90cd94395b8ca5c824"} Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.972023 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7d87d3642884141e5b7fbc72064aea61002a4b0453d8c90cd94395b8ca5c824" Oct 09 08:35:15 crc kubenswrapper[4872]: I1009 08:35:15.972074 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-wwgjr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217091 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:16 crc kubenswrapper[4872]: E1009 08:35:16.217487 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217512 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: E1009 08:35:16.217533 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" containerName="keystone-db-sync" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217541 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" containerName="keystone-db-sync" Oct 09 08:35:16 crc kubenswrapper[4872]: E1009 08:35:16.217562 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6154eba-fb06-4583-8937-9c1183083fec" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217572 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6154eba-fb06-4583-8937-9c1183083fec" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: E1009 08:35:16.217590 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3032b0c-ddae-450b-be63-1cd4b9014afb" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217599 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3032b0c-ddae-450b-be63-1cd4b9014afb" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217834 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217858 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3032b0c-ddae-450b-be63-1cd4b9014afb" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217874 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" containerName="keystone-db-sync" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.217890 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6154eba-fb06-4583-8937-9c1183083fec" containerName="mariadb-account-create" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.218870 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.242931 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.277246 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-xdjqg"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.278537 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.282199 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.282454 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.282586 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.284830 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kvgxv" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.301350 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xdjqg"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356660 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-42wc5\" (UniqueName: \"kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356716 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356737 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356757 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356776 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356800 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356816 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356851 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356871 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356906 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356949 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.356968 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc47t\" (UniqueName: \"kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.415889 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.435017 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.438720 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-2vfjh" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.438923 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.438953 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.439159 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.446526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461391 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461530 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461572 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc47t\" (UniqueName: \"kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461724 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-42wc5\" (UniqueName: \"kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461774 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461805 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461829 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461847 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461878 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461947 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.461984 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.464907 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.470142 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.475183 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.478099 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.478336 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.478852 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.479491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.486318 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.501350 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.507747 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.518517 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc47t\" (UniqueName: \"kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t\") pod \"keystone-bootstrap-xdjqg\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.519142 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.521570 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.524682 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-42wc5\" (UniqueName: \"kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5\") pod \"dnsmasq-dns-6f8c45789f-xdgdr\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.526808 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.537118 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.546901 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.547153 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565255 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565352 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565398 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565432 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565471 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565489 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565546 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565587 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565632 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2sj8\" (UniqueName: \"kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565705 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.565938 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.566090 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.574833 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zprz2"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.576414 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.582842 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.583272 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-h4hr6" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.583560 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.603264 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.604302 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.624423 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zprz2"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666621 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666678 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666697 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2sj8\" (UniqueName: \"kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666715 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666754 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666769 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666822 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666839 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666870 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666890 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666906 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666924 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k86n\" (UniqueName: \"kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666950 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666966 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.666982 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.667036 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.668386 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.671458 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.672205 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.672959 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.675166 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.688094 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.689277 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.689522 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.694715 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.694723 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.699209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.704599 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2sj8\" (UniqueName: \"kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.736097 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.752339 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " pod="openstack/ceilometer-0" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.765402 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd\") pod \"horizon-648775cb6f-77vfj\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.768940 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.775349 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.775619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k86n\" (UniqueName: \"kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.776112 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.776698 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.776782 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.776916 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777468 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777528 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777671 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777744 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777801 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzq8d\" (UniqueName: \"kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.777855 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.786113 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.791563 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.796460 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k86n\" (UniqueName: \"kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.797308 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.798059 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.812066 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data\") pod \"placement-db-sync-zprz2\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879379 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879422 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879443 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879471 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879497 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n89n\" (UniqueName: \"kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879518 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879535 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879562 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879584 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzq8d\" (UniqueName: \"kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.879623 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.881228 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.887288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.889542 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.890457 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.903088 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.906793 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzq8d\" (UniqueName: \"kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d\") pod \"dnsmasq-dns-fcfdd6f9f-gf42z\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.980576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.980624 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.980720 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n89n\" (UniqueName: \"kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.980745 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.980790 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.981572 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.981929 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.983759 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:16 crc kubenswrapper[4872]: I1009 08:35:16.985371 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.002900 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n89n\" (UniqueName: \"kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n\") pod \"horizon-66cb5748f-phdrk\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.052710 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.064619 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.069361 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-8jxqn"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.071024 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.071096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.080143 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kwdcb" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.080847 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.081660 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082009 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082061 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082087 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082121 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082196 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd598\" (UniqueName: \"kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.082588 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.093270 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8jxqn"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.114957 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.136920 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.184784 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.185066 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd598\" (UniqueName: \"kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.185186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.185757 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.185851 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.185922 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.186075 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.192038 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.192204 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.192405 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.194140 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.204061 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-xdjqg"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.210453 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd598\" (UniqueName: \"kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598\") pod \"cinder-db-sync-8jxqn\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.277308 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.374441 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-j52s6"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.375880 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.381389 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hml5z" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.381668 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.385268 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-gbw26"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.386498 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394326 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs57t\" (UniqueName: \"kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394375 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394443 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh7rb\" (UniqueName: \"kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394462 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394491 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394518 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.394860 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dpjfl" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.395045 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.395173 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.395795 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.397999 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j52s6"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.405811 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gbw26"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.495962 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs57t\" (UniqueName: \"kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.496032 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.496140 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh7rb\" (UniqueName: \"kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.496161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.496200 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.496226 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.503476 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.503608 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.509754 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.530926 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh7rb\" (UniqueName: \"kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.535997 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle\") pod \"neutron-db-sync-gbw26\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.539583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs57t\" (UniqueName: \"kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t\") pod \"barbican-db-sync-j52s6\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.612040 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:35:17 crc kubenswrapper[4872]: W1009 08:35:17.631837 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda4ba9d6d_2798_4832_9260_e055e35818b1.slice/crio-b8851cdb126658499d04225e64dc6643a88d34f356b3e27164f8ec6b80ba7580 WatchSource:0}: Error finding container b8851cdb126658499d04225e64dc6643a88d34f356b3e27164f8ec6b80ba7580: Status 404 returned error can't find the container with id b8851cdb126658499d04225e64dc6643a88d34f356b3e27164f8ec6b80ba7580 Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.679705 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.751236 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j52s6" Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.875010 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.890275 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zprz2"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.898141 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-8jxqn"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.912600 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.994363 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerStarted","Data":"59272a8444a79613e701d551f071a192fda2e2b4e9981c9939b3a7f8c4a446d0"} Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.996578 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdjqg" event={"ID":"177ee0b4-5eda-48c8-be97-a86b6f81e330","Type":"ContainerStarted","Data":"2744602f9826eba65667e9ba0edc1788b8a741e19c0e765e42b227cbede4f2e9"} Oct 09 08:35:17 crc kubenswrapper[4872]: I1009 08:35:17.996673 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdjqg" event={"ID":"177ee0b4-5eda-48c8-be97-a86b6f81e330","Type":"ContainerStarted","Data":"ef264c12b236b9df3682e934c638c721457322d85db723d9e064d088d77be8d5"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.000828 4872 generic.go:334] "Generic (PLEG): container finished" podID="25572511-a964-4ef5-accf-6a780c16f837" containerID="629150188dd06f497b769a6606f4ddc191b9480b0161d81703fd93e619447551" exitCode=0 Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.000886 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" event={"ID":"25572511-a964-4ef5-accf-6a780c16f837","Type":"ContainerDied","Data":"629150188dd06f497b769a6606f4ddc191b9480b0161d81703fd93e619447551"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.000908 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" event={"ID":"25572511-a964-4ef5-accf-6a780c16f837","Type":"ContainerStarted","Data":"d60c02723feb5485d3465d6d58164db9323397d85d1986fa6ed77196de310ff9"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.013008 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerStarted","Data":"d9f858e3fdcc797ab275aff901416be678fd2205f6be672d53d8cd55d0af019a"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.018087 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerStarted","Data":"b8851cdb126658499d04225e64dc6643a88d34f356b3e27164f8ec6b80ba7580"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.019587 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zprz2" event={"ID":"1376b2ed-ee38-4201-bfd3-f35ef4093fa0","Type":"ContainerStarted","Data":"bd6b7b873bedcccdad4afd05b5ac8e97d22d6109e027e47264a9c955a7d26bc7"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.020275 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.021169 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8jxqn" event={"ID":"88cf1623-86ec-41eb-802f-c0996b7442be","Type":"ContainerStarted","Data":"87a95a7545445b96d64fd264e33d71f8ed355bff740eab2a796877f4675bc0f3"} Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.029749 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-xdjqg" podStartSLOduration=2.029728183 podStartE2EDuration="2.029728183s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:18.013162755 +0000 UTC m=+1076.203691381" watchObservedRunningTime="2025-10-09 08:35:18.029728183 +0000 UTC m=+1076.220256809" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.199123 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-gbw26"] Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.291620 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-j52s6"] Oct 09 08:35:18 crc kubenswrapper[4872]: W1009 08:35:18.354624 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod658ddbb1_1aaf_4e34_9a59_c6d01a87c505.slice/crio-f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76 WatchSource:0}: Error finding container f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76: Status 404 returned error can't find the container with id f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76 Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.363714 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.419588 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.419923 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-42wc5\" (UniqueName: \"kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.420113 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.420181 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.420203 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.420453 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0\") pod \"25572511-a964-4ef5-accf-6a780c16f837\" (UID: \"25572511-a964-4ef5-accf-6a780c16f837\") " Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.425734 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5" (OuterVolumeSpecName: "kube-api-access-42wc5") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "kube-api-access-42wc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.504033 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config" (OuterVolumeSpecName: "config") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.522216 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.522242 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-42wc5\" (UniqueName: \"kubernetes.io/projected/25572511-a964-4ef5-accf-6a780c16f837-kube-api-access-42wc5\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.537690 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.545611 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.546410 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.550344 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "25572511-a964-4ef5-accf-6a780c16f837" (UID: "25572511-a964-4ef5-accf-6a780c16f837"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.628376 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.628417 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.628429 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.628439 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/25572511-a964-4ef5-accf-6a780c16f837-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.633055 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.642888 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:35:18 crc kubenswrapper[4872]: E1009 08:35:18.643272 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25572511-a964-4ef5-accf-6a780c16f837" containerName="init" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.643284 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25572511-a964-4ef5-accf-6a780c16f837" containerName="init" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.643466 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25572511-a964-4ef5-accf-6a780c16f837" containerName="init" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.644340 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.659492 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.704296 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.730522 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.730949 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.731067 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.731099 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.731125 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbljx\" (UniqueName: \"kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.832570 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.832622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbljx\" (UniqueName: \"kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.832692 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.832727 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.832811 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.834610 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.838730 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.838858 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.838914 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.865303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbljx\" (UniqueName: \"kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx\") pod \"horizon-5749cd8d8c-b6qs4\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:18 crc kubenswrapper[4872]: I1009 08:35:18.992140 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.041584 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gbw26" event={"ID":"3b8d1d3f-8d6e-458f-8ebe-da41817d436c","Type":"ContainerStarted","Data":"f8f4f178e43c02d36259b6374f52177a6612221341d63c993728d95100aa152d"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.041670 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gbw26" event={"ID":"3b8d1d3f-8d6e-458f-8ebe-da41817d436c","Type":"ContainerStarted","Data":"1c2b5c9dfa412be89c17dfbe7f40c3c38cd530255bb5038fc6a8473be03ddfe5"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.047222 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" event={"ID":"25572511-a964-4ef5-accf-6a780c16f837","Type":"ContainerDied","Data":"d60c02723feb5485d3465d6d58164db9323397d85d1986fa6ed77196de310ff9"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.047276 4872 scope.go:117] "RemoveContainer" containerID="629150188dd06f497b769a6606f4ddc191b9480b0161d81703fd93e619447551" Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.047593 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f8c45789f-xdgdr" Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.064617 4872 generic.go:334] "Generic (PLEG): container finished" podID="bf990753-621d-4d8b-b320-2540d3dc722c" containerID="15af2b37f0c78a931959ac54657df5857e1e3fdd70618ebb77e7d2e59d5a0243" exitCode=0 Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.064680 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" event={"ID":"bf990753-621d-4d8b-b320-2540d3dc722c","Type":"ContainerDied","Data":"15af2b37f0c78a931959ac54657df5857e1e3fdd70618ebb77e7d2e59d5a0243"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.064727 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" event={"ID":"bf990753-621d-4d8b-b320-2540d3dc722c","Type":"ContainerStarted","Data":"0a5cb03ae1b23e768fbf7b63995d87c18ed72c6883a689dafb709183c61c05c6"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.064757 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-gbw26" podStartSLOduration=2.064734771 podStartE2EDuration="2.064734771s" podCreationTimestamp="2025-10-09 08:35:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:19.058189942 +0000 UTC m=+1077.248718588" watchObservedRunningTime="2025-10-09 08:35:19.064734771 +0000 UTC m=+1077.255263417" Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.068976 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j52s6" event={"ID":"658ddbb1-1aaf-4e34-9a59-c6d01a87c505","Type":"ContainerStarted","Data":"f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76"} Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.197780 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.231686 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f8c45789f-xdgdr"] Oct 09 08:35:19 crc kubenswrapper[4872]: I1009 08:35:19.621696 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:35:20 crc kubenswrapper[4872]: I1009 08:35:20.088127 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerStarted","Data":"1efa65c66da0c21a141f5ac90fbf7670440bf7b5b5b8593409e9b20a33caa83f"} Oct 09 08:35:20 crc kubenswrapper[4872]: I1009 08:35:20.094052 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" event={"ID":"bf990753-621d-4d8b-b320-2540d3dc722c","Type":"ContainerStarted","Data":"116426969a0cf57d52722140e2e2d90d9cb4d2a6fc17db08e75e63c689a96d8e"} Oct 09 08:35:20 crc kubenswrapper[4872]: I1009 08:35:20.094299 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:20 crc kubenswrapper[4872]: I1009 08:35:20.122423 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" podStartSLOduration=4.122392924 podStartE2EDuration="4.122392924s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:20.115187187 +0000 UTC m=+1078.305715833" watchObservedRunningTime="2025-10-09 08:35:20.122392924 +0000 UTC m=+1078.312921550" Oct 09 08:35:20 crc kubenswrapper[4872]: I1009 08:35:20.498870 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25572511-a964-4ef5-accf-6a780c16f837" path="/var/lib/kubelet/pods/25572511-a964-4ef5-accf-6a780c16f837/volumes" Oct 09 08:35:22 crc kubenswrapper[4872]: I1009 08:35:22.120967 4872 generic.go:334] "Generic (PLEG): container finished" podID="256fd449-62bd-44fe-945e-92663ab62698" containerID="fc5356d5eb0b02f29e7d0e03b74c7bf865150c941377f1311308f3be7e10c087" exitCode=0 Oct 09 08:35:22 crc kubenswrapper[4872]: I1009 08:35:22.121054 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvtzz" event={"ID":"256fd449-62bd-44fe-945e-92663ab62698","Type":"ContainerDied","Data":"fc5356d5eb0b02f29e7d0e03b74c7bf865150c941377f1311308f3be7e10c087"} Oct 09 08:35:22 crc kubenswrapper[4872]: I1009 08:35:22.123740 4872 generic.go:334] "Generic (PLEG): container finished" podID="177ee0b4-5eda-48c8-be97-a86b6f81e330" containerID="2744602f9826eba65667e9ba0edc1788b8a741e19c0e765e42b227cbede4f2e9" exitCode=0 Oct 09 08:35:22 crc kubenswrapper[4872]: I1009 08:35:22.123778 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdjqg" event={"ID":"177ee0b4-5eda-48c8-be97-a86b6f81e330","Type":"ContainerDied","Data":"2744602f9826eba65667e9ba0edc1788b8a741e19c0e765e42b227cbede4f2e9"} Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.302598 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.341837 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.343447 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.346294 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.355442 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.399971 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.403991 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404053 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404084 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404108 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc6n4\" (UniqueName: \"kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404250 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404335 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.404391 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.443177 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5bf66bd496-2sdwc"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.444673 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.460023 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bf66bd496-2sdwc"] Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505616 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-tls-certs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505731 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-scripts\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505845 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505874 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-config-data\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505922 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505941 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvbt5\" (UniqueName: \"kubernetes.io/projected/6d37cb10-16dd-4360-aaf9-b90839973688-kube-api-access-kvbt5\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505964 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d37cb10-16dd-4360-aaf9-b90839973688-logs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.505991 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506016 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fc6n4\" (UniqueName: \"kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506088 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-secret-key\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506623 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-combined-ca-bundle\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506721 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506774 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.506799 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.507091 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.507459 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.513734 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.513787 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.513915 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.520591 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.525451 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fc6n4\" (UniqueName: \"kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4\") pod \"horizon-6cdd4b976-7cgml\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609672 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-tls-certs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609760 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-scripts\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609818 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-config-data\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609851 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvbt5\" (UniqueName: \"kubernetes.io/projected/6d37cb10-16dd-4360-aaf9-b90839973688-kube-api-access-kvbt5\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609869 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d37cb10-16dd-4360-aaf9-b90839973688-logs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609893 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-secret-key\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.609917 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-combined-ca-bundle\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.611417 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6d37cb10-16dd-4360-aaf9-b90839973688-logs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.612298 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-scripts\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.612599 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6d37cb10-16dd-4360-aaf9-b90839973688-config-data\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.614333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-tls-certs\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.615134 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-combined-ca-bundle\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.619504 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6d37cb10-16dd-4360-aaf9-b90839973688-horizon-secret-key\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.626497 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvbt5\" (UniqueName: \"kubernetes.io/projected/6d37cb10-16dd-4360-aaf9-b90839973688-kube-api-access-kvbt5\") pod \"horizon-5bf66bd496-2sdwc\" (UID: \"6d37cb10-16dd-4360-aaf9-b90839973688\") " pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.682167 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:25 crc kubenswrapper[4872]: I1009 08:35:25.768468 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:27 crc kubenswrapper[4872]: I1009 08:35:27.116792 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:27 crc kubenswrapper[4872]: I1009 08:35:27.175844 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:35:27 crc kubenswrapper[4872]: I1009 08:35:27.176117 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" containerID="cri-o://657b5aa5454781db9f6d0c43745909a179aa686d8bf45541b653a6696f7b3b42" gracePeriod=10 Oct 09 08:35:28 crc kubenswrapper[4872]: I1009 08:35:28.201275 4872 generic.go:334] "Generic (PLEG): container finished" podID="137de144-afdf-4553-be18-9a6e37243661" containerID="657b5aa5454781db9f6d0c43745909a179aa686d8bf45541b653a6696f7b3b42" exitCode=0 Oct 09 08:35:28 crc kubenswrapper[4872]: I1009 08:35:28.201323 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" event={"ID":"137de144-afdf-4553-be18-9a6e37243661","Type":"ContainerDied","Data":"657b5aa5454781db9f6d0c43745909a179aa686d8bf45541b653a6696f7b3b42"} Oct 09 08:35:28 crc kubenswrapper[4872]: I1009 08:35:28.276847 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: connect: connection refused" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.190430 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.228040 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-xdjqg" event={"ID":"177ee0b4-5eda-48c8-be97-a86b6f81e330","Type":"ContainerDied","Data":"ef264c12b236b9df3682e934c638c721457322d85db723d9e064d088d77be8d5"} Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.228119 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ef264c12b236b9df3682e934c638c721457322d85db723d9e064d088d77be8d5" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.228076 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-xdjqg" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.298432 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.298544 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.307053 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.307086 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.307202 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.307290 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc47t\" (UniqueName: \"kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t\") pod \"177ee0b4-5eda-48c8-be97-a86b6f81e330\" (UID: \"177ee0b4-5eda-48c8-be97-a86b6f81e330\") " Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.309051 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.310390 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts" (OuterVolumeSpecName: "scripts") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.311006 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.312375 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t" (OuterVolumeSpecName: "kube-api-access-fc47t") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "kube-api-access-fc47t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.340684 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.372813 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data" (OuterVolumeSpecName: "config-data") pod "177ee0b4-5eda-48c8-be97-a86b6f81e330" (UID: "177ee0b4-5eda-48c8-be97-a86b6f81e330"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413517 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413563 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc47t\" (UniqueName: \"kubernetes.io/projected/177ee0b4-5eda-48c8-be97-a86b6f81e330-kube-api-access-fc47t\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413585 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413596 4872 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413607 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:29 crc kubenswrapper[4872]: I1009 08:35:29.413617 4872 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/177ee0b4-5eda-48c8-be97-a86b6f81e330-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.299444 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-xdjqg"] Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.310173 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-xdjqg"] Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.391375 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-b9jgn"] Oct 09 08:35:30 crc kubenswrapper[4872]: E1009 08:35:30.391899 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="177ee0b4-5eda-48c8-be97-a86b6f81e330" containerName="keystone-bootstrap" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.391922 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="177ee0b4-5eda-48c8-be97-a86b6f81e330" containerName="keystone-bootstrap" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.392127 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="177ee0b4-5eda-48c8-be97-a86b6f81e330" containerName="keystone-bootstrap" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.393018 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.395886 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.395986 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.396134 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-kvgxv" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.396486 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.400071 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b9jgn"] Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434568 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434636 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434752 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434843 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l8qs\" (UniqueName: \"kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434880 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.434923 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.472224 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="177ee0b4-5eda-48c8-be97-a86b6f81e330" path="/var/lib/kubelet/pods/177ee0b4-5eda-48c8-be97-a86b6f81e330/volumes" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536552 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536676 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536759 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l8qs\" (UniqueName: \"kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536831 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.536868 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.540446 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.540863 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.541010 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.541196 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.547305 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.566680 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l8qs\" (UniqueName: \"kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs\") pod \"keystone-bootstrap-b9jgn\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:30 crc kubenswrapper[4872]: I1009 08:35:30.716963 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:31 crc kubenswrapper[4872]: E1009 08:35:31.363391 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Oct 09 08:35:31 crc kubenswrapper[4872]: E1009 08:35:31.363550 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9k86n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-zprz2_openstack(1376b2ed-ee38-4201-bfd3-f35ef4093fa0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 08:35:31 crc kubenswrapper[4872]: E1009 08:35:31.364704 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-zprz2" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" Oct 09 08:35:32 crc kubenswrapper[4872]: E1009 08:35:32.274722 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-zprz2" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.236916 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvtzz" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.281777 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2chm\" (UniqueName: \"kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm\") pod \"256fd449-62bd-44fe-945e-92663ab62698\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.281881 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data\") pod \"256fd449-62bd-44fe-945e-92663ab62698\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.281981 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle\") pod \"256fd449-62bd-44fe-945e-92663ab62698\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.282076 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-jvtzz" event={"ID":"256fd449-62bd-44fe-945e-92663ab62698","Type":"ContainerDied","Data":"2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391"} Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.282112 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2374c3f43bf03a3f6fc88bbfbec3d1e172a6059692a6abe0ea5b89bae2b51391" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.282173 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-jvtzz" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.282320 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data\") pod \"256fd449-62bd-44fe-945e-92663ab62698\" (UID: \"256fd449-62bd-44fe-945e-92663ab62698\") " Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.287722 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "256fd449-62bd-44fe-945e-92663ab62698" (UID: "256fd449-62bd-44fe-945e-92663ab62698"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.287774 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm" (OuterVolumeSpecName: "kube-api-access-d2chm") pod "256fd449-62bd-44fe-945e-92663ab62698" (UID: "256fd449-62bd-44fe-945e-92663ab62698"). InnerVolumeSpecName "kube-api-access-d2chm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.321240 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "256fd449-62bd-44fe-945e-92663ab62698" (UID: "256fd449-62bd-44fe-945e-92663ab62698"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.329417 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data" (OuterVolumeSpecName: "config-data") pod "256fd449-62bd-44fe-945e-92663ab62698" (UID: "256fd449-62bd-44fe-945e-92663ab62698"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.385112 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.385153 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.385166 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2chm\" (UniqueName: \"kubernetes.io/projected/256fd449-62bd-44fe-945e-92663ab62698-kube-api-access-d2chm\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:33 crc kubenswrapper[4872]: I1009 08:35:33.385180 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/256fd449-62bd-44fe-945e-92663ab62698-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.622860 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:35:34 crc kubenswrapper[4872]: E1009 08:35:34.623483 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="256fd449-62bd-44fe-945e-92663ab62698" containerName="glance-db-sync" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.623496 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="256fd449-62bd-44fe-945e-92663ab62698" containerName="glance-db-sync" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.623707 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="256fd449-62bd-44fe-945e-92663ab62698" containerName="glance-db-sync" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.628398 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.644632 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705365 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4dcs\" (UniqueName: \"kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705414 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705448 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705473 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.705501 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807684 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4dcs\" (UniqueName: \"kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807774 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807829 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807862 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807893 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.807941 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.809161 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.810591 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.810669 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.810758 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.811263 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.834848 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4dcs\" (UniqueName: \"kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs\") pod \"dnsmasq-dns-57c957c4ff-bvnfw\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:34 crc kubenswrapper[4872]: I1009 08:35:34.953543 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.559820 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.561604 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.563950 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-sk5xf" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.564077 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.564399 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.572147 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721370 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721434 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721462 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721478 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721510 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721541 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.721609 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2bs7r\" (UniqueName: \"kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.814337 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.815921 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.818172 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.822575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.822784 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.822927 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2bs7r\" (UniqueName: \"kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.823018 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.823117 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.823203 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.823279 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.824358 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.824482 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.824483 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.826006 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.833339 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.843885 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.851257 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.853098 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2bs7r\" (UniqueName: \"kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.877709 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.896260 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925224 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rkzh\" (UniqueName: \"kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925275 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925342 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925365 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925442 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925478 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:35 crc kubenswrapper[4872]: I1009 08:35:35.925561 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.027512 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rkzh\" (UniqueName: \"kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.027989 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.028780 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031385 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031470 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031628 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031848 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.031958 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.032336 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.040496 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.041032 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.045474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.050729 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rkzh\" (UniqueName: \"kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.086481 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.229515 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.741811 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:36 crc kubenswrapper[4872]: I1009 08:35:36.806660 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:38 crc kubenswrapper[4872]: I1009 08:35:38.157433 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:35:38 crc kubenswrapper[4872]: I1009 08:35:38.158764 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:35:38 crc kubenswrapper[4872]: I1009 08:35:38.275898 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.170345 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.170513 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sd598,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-8jxqn_openstack(88cf1623-86ec-41eb-802f-c0996b7442be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.171767 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-8jxqn" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.225509 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377123 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lszv\" (UniqueName: \"kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377204 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377238 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377414 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377484 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.377534 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc\") pod \"137de144-afdf-4553-be18-9a6e37243661\" (UID: \"137de144-afdf-4553-be18-9a6e37243661\") " Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.383902 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.383895 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" event={"ID":"137de144-afdf-4553-be18-9a6e37243661","Type":"ContainerDied","Data":"5b59872e86d6a7e5493b628eaa1aec4f4291d92cf65fe41a8c7d59a1513b546d"} Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.384328 4872 scope.go:117] "RemoveContainer" containerID="657b5aa5454781db9f6d0c43745909a179aa686d8bf45541b653a6696f7b3b42" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.385230 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-8jxqn" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.385426 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv" (OuterVolumeSpecName: "kube-api-access-9lszv") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "kube-api-access-9lszv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.436491 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.454606 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config" (OuterVolumeSpecName: "config") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.455043 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.455102 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.457362 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "137de144-afdf-4553-be18-9a6e37243661" (UID: "137de144-afdf-4553-be18-9a6e37243661"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480839 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lszv\" (UniqueName: \"kubernetes.io/projected/137de144-afdf-4553-be18-9a6e37243661-kube-api-access-9lszv\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480876 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480885 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480894 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480905 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.480914 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/137de144-afdf-4553-be18-9a6e37243661-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.710540 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.724294 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6d5b6d6b67-hlcbg"] Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.786083 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.786265 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xs57t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-j52s6_openstack(658ddbb1-1aaf-4e34-9a59-c6d01a87c505): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 08:35:42 crc kubenswrapper[4872]: E1009 08:35:42.787562 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-j52s6" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" Oct 09 08:35:42 crc kubenswrapper[4872]: I1009 08:35:42.904510 4872 scope.go:117] "RemoveContainer" containerID="0cae0539d236193ce2d98f4eec1c06d55908d89ac56192dfcf07ad748919b750" Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.276208 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6d5b6d6b67-hlcbg" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.123:5353: i/o timeout" Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.386822 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:35:43 crc kubenswrapper[4872]: W1009 08:35:43.407892 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d37cb10_16dd_4360_aaf9_b90839973688.slice/crio-13048b4c48ef681d765ebaef00504893d96c35ef6167506b95ec176b3fe45659 WatchSource:0}: Error finding container 13048b4c48ef681d765ebaef00504893d96c35ef6167506b95ec176b3fe45659: Status 404 returned error can't find the container with id 13048b4c48ef681d765ebaef00504893d96c35ef6167506b95ec176b3fe45659 Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.417561 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5bf66bd496-2sdwc"] Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.419961 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerStarted","Data":"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b"} Oct 09 08:35:43 crc kubenswrapper[4872]: E1009 08:35:43.422102 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-j52s6" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.682827 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-b9jgn"] Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.750086 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:43 crc kubenswrapper[4872]: I1009 08:35:43.872546 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:35:43 crc kubenswrapper[4872]: W1009 08:35:43.877159 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7de46c6b_64ae_48c0_9485_5c2c8685e8b9.slice/crio-ba4146edbfff7b3425a0d3b762411018e2baa2e5ec88ee2bdfe33c69a78b0460 WatchSource:0}: Error finding container ba4146edbfff7b3425a0d3b762411018e2baa2e5ec88ee2bdfe33c69a78b0460: Status 404 returned error can't find the container with id ba4146edbfff7b3425a0d3b762411018e2baa2e5ec88ee2bdfe33c69a78b0460 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.436829 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerStarted","Data":"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.437249 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerStarted","Data":"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.437269 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerStarted","Data":"782279b8f5b17fac627713657aea82da644534dcc851c1804e12f693d5fc92a2"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.454889 4872 generic.go:334] "Generic (PLEG): container finished" podID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerID="547420e4687f983d560790af1d1488f17c5bde8571e7949ecedb24e7bb806350" exitCode=0 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.455090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" event={"ID":"7de46c6b-64ae-48c0-9485-5c2c8685e8b9","Type":"ContainerDied","Data":"547420e4687f983d560790af1d1488f17c5bde8571e7949ecedb24e7bb806350"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.455374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" event={"ID":"7de46c6b-64ae-48c0-9485-5c2c8685e8b9","Type":"ContainerStarted","Data":"ba4146edbfff7b3425a0d3b762411018e2baa2e5ec88ee2bdfe33c69a78b0460"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.461952 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cdd4b976-7cgml" podStartSLOduration=19.461936199 podStartE2EDuration="19.461936199s" podCreationTimestamp="2025-10-09 08:35:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:44.458670595 +0000 UTC m=+1102.649199231" watchObservedRunningTime="2025-10-09 08:35:44.461936199 +0000 UTC m=+1102.652464825" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489341 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="137de144-afdf-4553-be18-9a6e37243661" path="/var/lib/kubelet/pods/137de144-afdf-4553-be18-9a6e37243661/volumes" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bf66bd496-2sdwc" event={"ID":"6d37cb10-16dd-4360-aaf9-b90839973688","Type":"ContainerStarted","Data":"9a8ef2c0001fa0a42ce3ec84761b771592eb51c7511a77ce2182dd689a46183c"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489961 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bf66bd496-2sdwc" event={"ID":"6d37cb10-16dd-4360-aaf9-b90839973688","Type":"ContainerStarted","Data":"1626350f9b7471e1aa8e5d2267a7f3d63185a6a74c7060f9b2c36f755b994941"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489970 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5bf66bd496-2sdwc" event={"ID":"6d37cb10-16dd-4360-aaf9-b90839973688","Type":"ContainerStarted","Data":"13048b4c48ef681d765ebaef00504893d96c35ef6167506b95ec176b3fe45659"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489981 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerStarted","Data":"bcd8456f97f41ab7e8fa72eea46f1098ba8acb3ea749bfeacfbbc46119e2e553"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.489990 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerStarted","Data":"5ac777c315e47e88e1d63e794729282c81a7d96d5b7b849cc598a8b0869f0d33"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.511556 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerStarted","Data":"bbd62b7f0c0a68e436020bd4fc66acbecd5e2cea57db6a55cdafe0ae317be93e"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.511628 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerStarted","Data":"0e540636c822a0b88202d76d69fbb28add2b27d431302b713682c776fe3fdb14"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.511856 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5749cd8d8c-b6qs4" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon-log" containerID="cri-o://0e540636c822a0b88202d76d69fbb28add2b27d431302b713682c776fe3fdb14" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.512313 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-5749cd8d8c-b6qs4" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon" containerID="cri-o://bbd62b7f0c0a68e436020bd4fc66acbecd5e2cea57db6a55cdafe0ae317be93e" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.531312 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b9jgn" event={"ID":"c6788083-19e2-4636-9c90-468ddecb1115","Type":"ContainerStarted","Data":"b5298ecc9b22f9956f047d1fc258e0013e0ade9f95007721882a668864ed57f7"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.531358 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b9jgn" event={"ID":"c6788083-19e2-4636-9c90-468ddecb1115","Type":"ContainerStarted","Data":"0f85545d672c6081808660f54094dac3e9aa475ce7ab4890bd09e41fc271415d"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.552553 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5bf66bd496-2sdwc" podStartSLOduration=19.552526663 podStartE2EDuration="19.552526663s" podCreationTimestamp="2025-10-09 08:35:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:44.515424992 +0000 UTC m=+1102.705953618" watchObservedRunningTime="2025-10-09 08:35:44.552526663 +0000 UTC m=+1102.743055289" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.561859 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerStarted","Data":"db797175d5e966c6e067219ff4a95c22c14a5f3bd5b25bc32545c9ea9c11554c"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.561897 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerStarted","Data":"672a9e02184b99f9a968b3657325a33f4757bc40a20b94f57965e7845b8d7a13"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.566710 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-648775cb6f-77vfj" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon-log" containerID="cri-o://672a9e02184b99f9a968b3657325a33f4757bc40a20b94f57965e7845b8d7a13" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.566915 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-648775cb6f-77vfj" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon" containerID="cri-o://db797175d5e966c6e067219ff4a95c22c14a5f3bd5b25bc32545c9ea9c11554c" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.581911 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerStarted","Data":"fa7e50b7bf9d20e414407cefeec9c0a3953becbf8defe0943a0370c0802cdafb"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.581968 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerStarted","Data":"401c2b173285cfab1735af4da1a043607f76c7e975f58eb132c70a3a0b12beaf"} Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.582029 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66cb5748f-phdrk" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon-log" containerID="cri-o://401c2b173285cfab1735af4da1a043607f76c7e975f58eb132c70a3a0b12beaf" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.582126 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66cb5748f-phdrk" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon" containerID="cri-o://fa7e50b7bf9d20e414407cefeec9c0a3953becbf8defe0943a0370c0802cdafb" gracePeriod=30 Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.612728 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.624106 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5749cd8d8c-b6qs4" podStartSLOduration=3.332441044 podStartE2EDuration="26.624082918s" podCreationTimestamp="2025-10-09 08:35:18 +0000 UTC" firstStartedPulling="2025-10-09 08:35:19.635428601 +0000 UTC m=+1077.825957217" lastFinishedPulling="2025-10-09 08:35:42.927070465 +0000 UTC m=+1101.117599091" observedRunningTime="2025-10-09 08:35:44.559250157 +0000 UTC m=+1102.749778803" watchObservedRunningTime="2025-10-09 08:35:44.624082918 +0000 UTC m=+1102.814611564" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.647177 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-b9jgn" podStartSLOduration=14.647151564 podStartE2EDuration="14.647151564s" podCreationTimestamp="2025-10-09 08:35:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:44.593754543 +0000 UTC m=+1102.784283189" watchObservedRunningTime="2025-10-09 08:35:44.647151564 +0000 UTC m=+1102.837680190" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.673931 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-648775cb6f-77vfj" podStartSLOduration=3.656481698 podStartE2EDuration="28.673913146s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="2025-10-09 08:35:17.892043419 +0000 UTC m=+1076.082572045" lastFinishedPulling="2025-10-09 08:35:42.909474867 +0000 UTC m=+1101.100003493" observedRunningTime="2025-10-09 08:35:44.616457308 +0000 UTC m=+1102.806985944" watchObservedRunningTime="2025-10-09 08:35:44.673913146 +0000 UTC m=+1102.864441772" Oct 09 08:35:44 crc kubenswrapper[4872]: I1009 08:35:44.694227 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66cb5748f-phdrk" podStartSLOduration=3.815647632 podStartE2EDuration="28.694206702s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="2025-10-09 08:35:17.930457148 +0000 UTC m=+1076.120985774" lastFinishedPulling="2025-10-09 08:35:42.809016218 +0000 UTC m=+1100.999544844" observedRunningTime="2025-10-09 08:35:44.640248975 +0000 UTC m=+1102.830777601" watchObservedRunningTime="2025-10-09 08:35:44.694206702 +0000 UTC m=+1102.884735348" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.596290 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" event={"ID":"7de46c6b-64ae-48c0-9485-5c2c8685e8b9","Type":"ContainerStarted","Data":"201aee04d487f4b7c89d007e33548199cf03d4ca4794849d8bc07b634a6c1d81"} Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.596613 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.598444 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerStarted","Data":"b00d691a78feeb75c818c9e6da12b5fce94f15f2e1a01eb009f89ffdb480c463"} Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.600732 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerStarted","Data":"dce5e5b0b0f2782d02c18f324b8c3a56ae44440a849738813925852f7f3d37eb"} Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.600787 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerStarted","Data":"2af931eb52bd22744d29a2850c2884b3b9a1426b0e46f26487492145c0b83f7f"} Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.624100 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" podStartSLOduration=11.624082728 podStartE2EDuration="11.624082728s" podCreationTimestamp="2025-10-09 08:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:45.620040031 +0000 UTC m=+1103.810568657" watchObservedRunningTime="2025-10-09 08:35:45.624082728 +0000 UTC m=+1103.814611354" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.682197 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.682245 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.769287 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:45 crc kubenswrapper[4872]: I1009 08:35:45.769345 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.613693 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerStarted","Data":"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e"} Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.615724 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerStarted","Data":"f0a4580a74aac6aeca707abe71a6c3e3412ca12ba56340733515caa6cf65c087"} Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.615851 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-log" containerID="cri-o://dce5e5b0b0f2782d02c18f324b8c3a56ae44440a849738813925852f7f3d37eb" gracePeriod=30 Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.615909 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-log" containerID="cri-o://bcd8456f97f41ab7e8fa72eea46f1098ba8acb3ea749bfeacfbbc46119e2e553" gracePeriod=30 Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.615976 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-httpd" containerID="cri-o://f0a4580a74aac6aeca707abe71a6c3e3412ca12ba56340733515caa6cf65c087" gracePeriod=30 Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.615996 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-httpd" containerID="cri-o://b00d691a78feeb75c818c9e6da12b5fce94f15f2e1a01eb009f89ffdb480c463" gracePeriod=30 Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.651850 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=12.651827237 podStartE2EDuration="12.651827237s" podCreationTimestamp="2025-10-09 08:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:46.643535578 +0000 UTC m=+1104.834064214" watchObservedRunningTime="2025-10-09 08:35:46.651827237 +0000 UTC m=+1104.842355863" Oct 09 08:35:46 crc kubenswrapper[4872]: I1009 08:35:46.678608 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=12.67858795 podStartE2EDuration="12.67858795s" podCreationTimestamp="2025-10-09 08:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:46.671707801 +0000 UTC m=+1104.862236427" watchObservedRunningTime="2025-10-09 08:35:46.67858795 +0000 UTC m=+1104.869116576" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.065966 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.137837 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.639722 4872 generic.go:334] "Generic (PLEG): container finished" podID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerID="b00d691a78feeb75c818c9e6da12b5fce94f15f2e1a01eb009f89ffdb480c463" exitCode=0 Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.639764 4872 generic.go:334] "Generic (PLEG): container finished" podID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerID="bcd8456f97f41ab7e8fa72eea46f1098ba8acb3ea749bfeacfbbc46119e2e553" exitCode=143 Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.639813 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerDied","Data":"b00d691a78feeb75c818c9e6da12b5fce94f15f2e1a01eb009f89ffdb480c463"} Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.639843 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerDied","Data":"bcd8456f97f41ab7e8fa72eea46f1098ba8acb3ea749bfeacfbbc46119e2e553"} Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653371 4872 generic.go:334] "Generic (PLEG): container finished" podID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerID="f0a4580a74aac6aeca707abe71a6c3e3412ca12ba56340733515caa6cf65c087" exitCode=0 Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653412 4872 generic.go:334] "Generic (PLEG): container finished" podID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerID="dce5e5b0b0f2782d02c18f324b8c3a56ae44440a849738813925852f7f3d37eb" exitCode=143 Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653435 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerDied","Data":"f0a4580a74aac6aeca707abe71a6c3e3412ca12ba56340733515caa6cf65c087"} Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653462 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerDied","Data":"dce5e5b0b0f2782d02c18f324b8c3a56ae44440a849738813925852f7f3d37eb"} Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653472 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bff9394e-c319-4a4b-a018-8e696cfe2f7c","Type":"ContainerDied","Data":"2af931eb52bd22744d29a2850c2884b3b9a1426b0e46f26487492145c0b83f7f"} Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.653483 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2af931eb52bd22744d29a2850c2884b3b9a1426b0e46f26487492145c0b83f7f" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.682515 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787347 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2bs7r\" (UniqueName: \"kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787410 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787452 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787603 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787655 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.787684 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts\") pod \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\" (UID: \"bff9394e-c319-4a4b-a018-8e696cfe2f7c\") " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.794729 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs" (OuterVolumeSpecName: "logs") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.794983 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.820704 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r" (OuterVolumeSpecName: "kube-api-access-2bs7r") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "kube-api-access-2bs7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.832810 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.835901 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts" (OuterVolumeSpecName: "scripts") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.889418 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2bs7r\" (UniqueName: \"kubernetes.io/projected/bff9394e-c319-4a4b-a018-8e696cfe2f7c-kube-api-access-2bs7r\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.889692 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.889785 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.889855 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bff9394e-c319-4a4b-a018-8e696cfe2f7c-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.889929 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.910784 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.933269 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 09 08:35:47 crc kubenswrapper[4872]: I1009 08:35:47.945768 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data" (OuterVolumeSpecName: "config-data") pod "bff9394e-c319-4a4b-a018-8e696cfe2f7c" (UID: "bff9394e-c319-4a4b-a018-8e696cfe2f7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:47.999803 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.000067 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bff9394e-c319-4a4b-a018-8e696cfe2f7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.000080 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.086399 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101270 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101406 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101508 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101560 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101577 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101925 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rkzh\" (UniqueName: \"kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.101982 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts\") pod \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\" (UID: \"5759b6a6-94af-4af3-8cbf-fa5a5772b162\") " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.102056 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs" (OuterVolumeSpecName: "logs") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.102090 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.102366 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.102377 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5759b6a6-94af-4af3-8cbf-fa5a5772b162-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.108525 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh" (OuterVolumeSpecName: "kube-api-access-9rkzh") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "kube-api-access-9rkzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.121389 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.121742 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts" (OuterVolumeSpecName: "scripts") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.141470 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.169244 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data" (OuterVolumeSpecName: "config-data") pod "5759b6a6-94af-4af3-8cbf-fa5a5772b162" (UID: "5759b6a6-94af-4af3-8cbf-fa5a5772b162"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.203406 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.203456 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.203468 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rkzh\" (UniqueName: \"kubernetes.io/projected/5759b6a6-94af-4af3-8cbf-fa5a5772b162-kube-api-access-9rkzh\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.203479 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.203487 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5759b6a6-94af-4af3-8cbf-fa5a5772b162-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.227846 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.306347 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.670157 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.670171 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5759b6a6-94af-4af3-8cbf-fa5a5772b162","Type":"ContainerDied","Data":"5ac777c315e47e88e1d63e794729282c81a7d96d5b7b849cc598a8b0869f0d33"} Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.670612 4872 scope.go:117] "RemoveContainer" containerID="b00d691a78feeb75c818c9e6da12b5fce94f15f2e1a01eb009f89ffdb480c463" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.672507 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.672793 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zprz2" event={"ID":"1376b2ed-ee38-4201-bfd3-f35ef4093fa0","Type":"ContainerStarted","Data":"9f6d857c409a7c3371ea3f396ae5b825dab62973f557011b10fb55d0bf935d99"} Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.697455 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zprz2" podStartSLOduration=2.445027418 podStartE2EDuration="32.697440341s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="2025-10-09 08:35:17.917175785 +0000 UTC m=+1076.107704401" lastFinishedPulling="2025-10-09 08:35:48.169588698 +0000 UTC m=+1106.360117324" observedRunningTime="2025-10-09 08:35:48.696479063 +0000 UTC m=+1106.887007689" watchObservedRunningTime="2025-10-09 08:35:48.697440341 +0000 UTC m=+1106.887968967" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.733536 4872 scope.go:117] "RemoveContainer" containerID="bcd8456f97f41ab7e8fa72eea46f1098ba8acb3ea749bfeacfbbc46119e2e553" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.778524 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.816132 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.847738 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.868719 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.873126 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874019 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874068 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874088 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874097 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874122 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874154 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874172 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="init" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874179 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="init" Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874190 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874197 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: E1009 08:35:48.874237 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874249 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874537 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="137de144-afdf-4553-be18-9a6e37243661" containerName="dnsmasq-dns" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874591 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874605 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-log" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874619 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.874681 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" containerName="glance-httpd" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.877189 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.879732 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.884056 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.884334 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.885892 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.886025 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.888050 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-sk5xf" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.889203 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.890882 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.894090 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.894286 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 09 08:35:48 crc kubenswrapper[4872]: I1009 08:35:48.993741 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026087 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026135 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026162 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026182 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026199 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026225 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026270 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026296 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026314 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026334 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b82gc\" (UniqueName: \"kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026355 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026378 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cgmw\" (UniqueName: \"kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026400 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026416 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026432 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.026451 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127569 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127623 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127681 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b82gc\" (UniqueName: \"kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127727 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127754 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cgmw\" (UniqueName: \"kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127781 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127794 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127809 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127830 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127862 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127882 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127905 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127925 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127941 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.127973 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.128031 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.128039 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.128355 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.128788 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.129176 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.134083 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.134291 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.135042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.137154 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.137163 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.137596 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.140793 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.143709 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.148233 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.155068 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b82gc\" (UniqueName: \"kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.161911 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.165557 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cgmw\" (UniqueName: \"kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.179985 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.207320 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.213040 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.239118 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.697818 4872 generic.go:334] "Generic (PLEG): container finished" podID="3b8d1d3f-8d6e-458f-8ebe-da41817d436c" containerID="f8f4f178e43c02d36259b6374f52177a6612221341d63c993728d95100aa152d" exitCode=0 Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.698027 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gbw26" event={"ID":"3b8d1d3f-8d6e-458f-8ebe-da41817d436c","Type":"ContainerDied","Data":"f8f4f178e43c02d36259b6374f52177a6612221341d63c993728d95100aa152d"} Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.721437 4872 generic.go:334] "Generic (PLEG): container finished" podID="c6788083-19e2-4636-9c90-468ddecb1115" containerID="b5298ecc9b22f9956f047d1fc258e0013e0ade9f95007721882a668864ed57f7" exitCode=0 Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.721590 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b9jgn" event={"ID":"c6788083-19e2-4636-9c90-468ddecb1115","Type":"ContainerDied","Data":"b5298ecc9b22f9956f047d1fc258e0013e0ade9f95007721882a668864ed57f7"} Oct 09 08:35:49 crc kubenswrapper[4872]: I1009 08:35:49.958844 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.022911 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.035120 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.035380 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="dnsmasq-dns" containerID="cri-o://116426969a0cf57d52722140e2e2d90d9cb4d2a6fc17db08e75e63c689a96d8e" gracePeriod=10 Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.151526 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.480146 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5759b6a6-94af-4af3-8cbf-fa5a5772b162" path="/var/lib/kubelet/pods/5759b6a6-94af-4af3-8cbf-fa5a5772b162/volumes" Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.481341 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bff9394e-c319-4a4b-a018-8e696cfe2f7c" path="/var/lib/kubelet/pods/bff9394e-c319-4a4b-a018-8e696cfe2f7c/volumes" Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.788952 4872 generic.go:334] "Generic (PLEG): container finished" podID="bf990753-621d-4d8b-b320-2540d3dc722c" containerID="116426969a0cf57d52722140e2e2d90d9cb4d2a6fc17db08e75e63c689a96d8e" exitCode=0 Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.789075 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" event={"ID":"bf990753-621d-4d8b-b320-2540d3dc722c","Type":"ContainerDied","Data":"116426969a0cf57d52722140e2e2d90d9cb4d2a6fc17db08e75e63c689a96d8e"} Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.789107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" event={"ID":"bf990753-621d-4d8b-b320-2540d3dc722c","Type":"ContainerDied","Data":"0a5cb03ae1b23e768fbf7b63995d87c18ed72c6883a689dafb709183c61c05c6"} Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.789120 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a5cb03ae1b23e768fbf7b63995d87c18ed72c6883a689dafb709183c61c05c6" Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.811326 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerStarted","Data":"57bc46d3d46ca86bd57cf6caf3485ce07f16cf88947295820b44cc3f1b2fa7be"} Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.812311 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:50 crc kubenswrapper[4872]: I1009 08:35:50.830943 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerStarted","Data":"4e8701d58d65de16e060f9fd2ed02e7222dc56842d04161220209481a7845e2f"} Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014191 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014350 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014374 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014411 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014456 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzq8d\" (UniqueName: \"kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.014512 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb\") pod \"bf990753-621d-4d8b-b320-2540d3dc722c\" (UID: \"bf990753-621d-4d8b-b320-2540d3dc722c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.051071 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d" (OuterVolumeSpecName: "kube-api-access-nzq8d") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "kube-api-access-nzq8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.084402 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config" (OuterVolumeSpecName: "config") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.129816 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzq8d\" (UniqueName: \"kubernetes.io/projected/bf990753-621d-4d8b-b320-2540d3dc722c-kube-api-access-nzq8d\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.129855 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.141194 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.166416 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.176864 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.177423 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bf990753-621d-4d8b-b320-2540d3dc722c" (UID: "bf990753-621d-4d8b-b320-2540d3dc722c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.230552 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.230582 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.230591 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.230599 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/bf990753-621d-4d8b-b320-2540d3dc722c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.373538 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.430932 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.541053 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6l8qs\" (UniqueName: \"kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.541717 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.541779 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh7rb\" (UniqueName: \"kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb\") pod \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.542970 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config\") pod \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.543013 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.543044 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle\") pod \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\" (UID: \"3b8d1d3f-8d6e-458f-8ebe-da41817d436c\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.543075 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.543159 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.543194 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys\") pod \"c6788083-19e2-4636-9c90-468ddecb1115\" (UID: \"c6788083-19e2-4636-9c90-468ddecb1115\") " Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.550162 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts" (OuterVolumeSpecName: "scripts") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.551850 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.559951 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs" (OuterVolumeSpecName: "kube-api-access-6l8qs") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "kube-api-access-6l8qs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.576855 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb" (OuterVolumeSpecName: "kube-api-access-kh7rb") pod "3b8d1d3f-8d6e-458f-8ebe-da41817d436c" (UID: "3b8d1d3f-8d6e-458f-8ebe-da41817d436c"). InnerVolumeSpecName "kube-api-access-kh7rb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.576944 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.578894 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b8d1d3f-8d6e-458f-8ebe-da41817d436c" (UID: "3b8d1d3f-8d6e-458f-8ebe-da41817d436c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.596164 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.611842 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config" (OuterVolumeSpecName: "config") pod "3b8d1d3f-8d6e-458f-8ebe-da41817d436c" (UID: "3b8d1d3f-8d6e-458f-8ebe-da41817d436c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.611935 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data" (OuterVolumeSpecName: "config-data") pod "c6788083-19e2-4636-9c90-468ddecb1115" (UID: "c6788083-19e2-4636-9c90-468ddecb1115"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645187 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6l8qs\" (UniqueName: \"kubernetes.io/projected/c6788083-19e2-4636-9c90-468ddecb1115-kube-api-access-6l8qs\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645224 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645236 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh7rb\" (UniqueName: \"kubernetes.io/projected/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-kube-api-access-kh7rb\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645247 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645258 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645268 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b8d1d3f-8d6e-458f-8ebe-da41817d436c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645281 4872 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645291 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.645300 4872 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c6788083-19e2-4636-9c90-468ddecb1115-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.858101 4872 generic.go:334] "Generic (PLEG): container finished" podID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" containerID="9f6d857c409a7c3371ea3f396ae5b825dab62973f557011b10fb55d0bf935d99" exitCode=0 Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.858179 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zprz2" event={"ID":"1376b2ed-ee38-4201-bfd3-f35ef4093fa0","Type":"ContainerDied","Data":"9f6d857c409a7c3371ea3f396ae5b825dab62973f557011b10fb55d0bf935d99"} Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.859958 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerStarted","Data":"95bb1ad59c73749dba34c8eb0bf532fc6ef770f520d61395aa50daf85ca92446"} Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.903135 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-gbw26" event={"ID":"3b8d1d3f-8d6e-458f-8ebe-da41817d436c","Type":"ContainerDied","Data":"1c2b5c9dfa412be89c17dfbe7f40c3c38cd530255bb5038fc6a8473be03ddfe5"} Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.903174 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c2b5c9dfa412be89c17dfbe7f40c3c38cd530255bb5038fc6a8473be03ddfe5" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.903796 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-gbw26" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.927577 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7489b48876-4zz4g"] Oct 09 08:35:51 crc kubenswrapper[4872]: E1009 08:35:51.927966 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="dnsmasq-dns" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.927989 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="dnsmasq-dns" Oct 09 08:35:51 crc kubenswrapper[4872]: E1009 08:35:51.928000 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6788083-19e2-4636-9c90-468ddecb1115" containerName="keystone-bootstrap" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928007 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6788083-19e2-4636-9c90-468ddecb1115" containerName="keystone-bootstrap" Oct 09 08:35:51 crc kubenswrapper[4872]: E1009 08:35:51.928016 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b8d1d3f-8d6e-458f-8ebe-da41817d436c" containerName="neutron-db-sync" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928028 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b8d1d3f-8d6e-458f-8ebe-da41817d436c" containerName="neutron-db-sync" Oct 09 08:35:51 crc kubenswrapper[4872]: E1009 08:35:51.928057 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="init" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928070 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="init" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928292 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6788083-19e2-4636-9c90-468ddecb1115" containerName="keystone-bootstrap" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928310 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" containerName="dnsmasq-dns" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.928325 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b8d1d3f-8d6e-458f-8ebe-da41817d436c" containerName="neutron-db-sync" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.929091 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.945728 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7489b48876-4zz4g"] Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.948005 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-b9jgn" event={"ID":"c6788083-19e2-4636-9c90-468ddecb1115","Type":"ContainerDied","Data":"0f85545d672c6081808660f54094dac3e9aa475ce7ab4890bd09e41fc271415d"} Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.948039 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f85545d672c6081808660f54094dac3e9aa475ce7ab4890bd09e41fc271415d" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.948112 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-b9jgn" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.949961 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.950066 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968118 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-combined-ca-bundle\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968196 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g85r\" (UniqueName: \"kubernetes.io/projected/7f746daf-5432-4790-961e-c02910ec5946-kube-api-access-4g85r\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968227 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-credential-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968339 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-internal-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968362 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-scripts\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-fernet-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968511 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-public-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.968541 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-config-data\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.975970 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fcfdd6f9f-gf42z" Oct 09 08:35:51 crc kubenswrapper[4872]: I1009 08:35:51.977705 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerStarted","Data":"42aac7f2fd68b309be6ee6a3668a30de61cf6ee3f575a160f926b70613f1ce8c"} Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.028914 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.030763 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072487 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-public-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072539 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-config-data\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-combined-ca-bundle\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072679 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g85r\" (UniqueName: \"kubernetes.io/projected/7f746daf-5432-4790-961e-c02910ec5946-kube-api-access-4g85r\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072707 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-credential-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072771 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-internal-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-scripts\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.072822 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-fernet-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.075946 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.085592 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-scripts\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.089834 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-internal-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.092967 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-config-data\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.097386 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-public-tls-certs\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.103357 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-fernet-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.104447 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-credential-keys\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.115382 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f746daf-5432-4790-961e-c02910ec5946-combined-ca-bundle\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.115463 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.117302 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.130028 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.130540 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.130855 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-dpjfl" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.131334 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.148423 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g85r\" (UniqueName: \"kubernetes.io/projected/7f746daf-5432-4790-961e-c02910ec5946-kube-api-access-4g85r\") pod \"keystone-7489b48876-4zz4g\" (UID: \"7f746daf-5432-4790-961e-c02910ec5946\") " pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.173948 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.174055 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.174110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4kq5\" (UniqueName: \"kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.174140 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.174176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.174210 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.184721 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.201876 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.217411 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-fcfdd6f9f-gf42z"] Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284700 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284775 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284850 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284868 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284886 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284906 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mn7p\" (UniqueName: \"kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284933 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284960 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4kq5\" (UniqueName: \"kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.284983 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.285010 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.285047 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.285799 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.285821 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.286029 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.286491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.286587 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.307960 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.320140 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4kq5\" (UniqueName: \"kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5\") pod \"dnsmasq-dns-5ccc5c4795-27zw7\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.387936 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.387997 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.389919 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.389953 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.390002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mn7p\" (UniqueName: \"kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.396811 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.397880 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.398025 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.409413 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mn7p\" (UniqueName: \"kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.410675 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs\") pod \"neutron-66fb896578-k7hvx\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.522240 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.523008 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf990753-621d-4d8b-b320-2540d3dc722c" path="/var/lib/kubelet/pods/bf990753-621d-4d8b-b320-2540d3dc722c/volumes" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.541634 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:35:52 crc kubenswrapper[4872]: I1009 08:35:52.834099 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7489b48876-4zz4g"] Oct 09 08:35:53 crc kubenswrapper[4872]: I1009 08:35:53.058845 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerStarted","Data":"fdf21793e2e99ec9ff1aabbb3d51e6b0ba343ff4ca8b086f7944229443fd5b88"} Oct 09 08:35:53 crc kubenswrapper[4872]: I1009 08:35:53.068562 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerStarted","Data":"ecdd549b08379bea6014ca0133872d82b31cc5ca73b37ca8701c2b32fa88291d"} Oct 09 08:35:53 crc kubenswrapper[4872]: I1009 08:35:53.113338 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.113311759 podStartE2EDuration="5.113311759s" podCreationTimestamp="2025-10-09 08:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:53.079013669 +0000 UTC m=+1111.269542295" watchObservedRunningTime="2025-10-09 08:35:53.113311759 +0000 UTC m=+1111.303840385" Oct 09 08:35:53 crc kubenswrapper[4872]: I1009 08:35:53.124989 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.124968965 podStartE2EDuration="5.124968965s" podCreationTimestamp="2025-10-09 08:35:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:35:53.105817022 +0000 UTC m=+1111.296345638" watchObservedRunningTime="2025-10-09 08:35:53.124968965 +0000 UTC m=+1111.315497611" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.392916 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c68d5ff89-tcgwk"] Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.396481 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.400000 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.400855 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.426587 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c68d5ff89-tcgwk"] Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570119 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-httpd-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-ovndb-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570277 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570348 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-internal-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570417 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-combined-ca-bundle\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570469 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-public-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.570527 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8rl4\" (UniqueName: \"kubernetes.io/projected/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-kube-api-access-q8rl4\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672128 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-public-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672198 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8rl4\" (UniqueName: \"kubernetes.io/projected/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-kube-api-access-q8rl4\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672313 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-httpd-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672364 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-ovndb-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672402 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672480 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-internal-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.672516 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-combined-ca-bundle\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.680284 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-httpd-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.683042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-public-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.684016 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-combined-ca-bundle\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.684573 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-ovndb-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.693385 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-config\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.696506 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8rl4\" (UniqueName: \"kubernetes.io/projected/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-kube-api-access-q8rl4\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.698454 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0-internal-tls-certs\") pod \"neutron-c68d5ff89-tcgwk\" (UID: \"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0\") " pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:54 crc kubenswrapper[4872]: I1009 08:35:54.743030 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:35:55 crc kubenswrapper[4872]: I1009 08:35:55.684205 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 09 08:35:55 crc kubenswrapper[4872]: I1009 08:35:55.772402 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5bf66bd496-2sdwc" podUID="6d37cb10-16dd-4360-aaf9-b90839973688" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 09 08:35:59 crc kubenswrapper[4872]: W1009 08:35:59.011965 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f746daf_5432_4790_961e_c02910ec5946.slice/crio-0c919d74adf3d76b7cbda7e12538cf9b800f40a6ae3446f12870fdc59a030e63 WatchSource:0}: Error finding container 0c919d74adf3d76b7cbda7e12538cf9b800f40a6ae3446f12870fdc59a030e63: Status 404 returned error can't find the container with id 0c919d74adf3d76b7cbda7e12538cf9b800f40a6ae3446f12870fdc59a030e63 Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.123931 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7489b48876-4zz4g" event={"ID":"7f746daf-5432-4790-961e-c02910ec5946","Type":"ContainerStarted","Data":"0c919d74adf3d76b7cbda7e12538cf9b800f40a6ae3446f12870fdc59a030e63"} Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.130316 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zprz2" event={"ID":"1376b2ed-ee38-4201-bfd3-f35ef4093fa0","Type":"ContainerDied","Data":"bd6b7b873bedcccdad4afd05b5ac8e97d22d6109e027e47264a9c955a7d26bc7"} Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.130365 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd6b7b873bedcccdad4afd05b5ac8e97d22d6109e027e47264a9c955a7d26bc7" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.214416 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.214854 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.240399 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.240456 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.283963 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.284026 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.326362 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zprz2" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.335907 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.352701 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.479030 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs\") pod \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.479108 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle\") pod \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.479179 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k86n\" (UniqueName: \"kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n\") pod \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.479241 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data\") pod \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.479284 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts\") pod \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\" (UID: \"1376b2ed-ee38-4201-bfd3-f35ef4093fa0\") " Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.480250 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs" (OuterVolumeSpecName: "logs") pod "1376b2ed-ee38-4201-bfd3-f35ef4093fa0" (UID: "1376b2ed-ee38-4201-bfd3-f35ef4093fa0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.483131 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.492223 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n" (OuterVolumeSpecName: "kube-api-access-9k86n") pod "1376b2ed-ee38-4201-bfd3-f35ef4093fa0" (UID: "1376b2ed-ee38-4201-bfd3-f35ef4093fa0"). InnerVolumeSpecName "kube-api-access-9k86n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.493831 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts" (OuterVolumeSpecName: "scripts") pod "1376b2ed-ee38-4201-bfd3-f35ef4093fa0" (UID: "1376b2ed-ee38-4201-bfd3-f35ef4093fa0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.572583 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1376b2ed-ee38-4201-bfd3-f35ef4093fa0" (UID: "1376b2ed-ee38-4201-bfd3-f35ef4093fa0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.576610 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data" (OuterVolumeSpecName: "config-data") pod "1376b2ed-ee38-4201-bfd3-f35ef4093fa0" (UID: "1376b2ed-ee38-4201-bfd3-f35ef4093fa0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.584560 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.584852 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k86n\" (UniqueName: \"kubernetes.io/projected/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-kube-api-access-9k86n\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.584937 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.585002 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1376b2ed-ee38-4201-bfd3-f35ef4093fa0-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.666974 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:35:59 crc kubenswrapper[4872]: I1009 08:35:59.912981 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.153296 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j52s6" event={"ID":"658ddbb1-1aaf-4e34-9a59-c6d01a87c505","Type":"ContainerStarted","Data":"b2f2219e4942c73b94cb762a64e5ee5f127a4f829ea5c155feeb42e570e44b3e"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.158805 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerStarted","Data":"4169a26ade0e61ba21af3d85ff236c1ca06bc83d97c70860032741af62520130"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.166539 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7489b48876-4zz4g" event={"ID":"7f746daf-5432-4790-961e-c02910ec5946","Type":"ContainerStarted","Data":"7091dfa81ed65d546bc6e74dbc5abb168dbe4b95b5e6e4573476f6cc608c8da5"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.166673 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.175297 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-j52s6" podStartSLOduration=1.8570841489999999 podStartE2EDuration="43.1752808s" podCreationTimestamp="2025-10-09 08:35:17 +0000 UTC" firstStartedPulling="2025-10-09 08:35:18.357390938 +0000 UTC m=+1076.547919554" lastFinishedPulling="2025-10-09 08:35:59.675587579 +0000 UTC m=+1117.866116205" observedRunningTime="2025-10-09 08:36:00.170714668 +0000 UTC m=+1118.361243304" watchObservedRunningTime="2025-10-09 08:36:00.1752808 +0000 UTC m=+1118.365809426" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.186568 4872 generic.go:334] "Generic (PLEG): container finished" podID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerID="e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee" exitCode=0 Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.186947 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" event={"ID":"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c","Type":"ContainerDied","Data":"e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.187067 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" event={"ID":"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c","Type":"ContainerStarted","Data":"b3cafa624f9ae998609735a16478372c05d9ade6977d3fee62dbfbd22c651a58"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.211512 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7489b48876-4zz4g" podStartSLOduration=9.211491255 podStartE2EDuration="9.211491255s" podCreationTimestamp="2025-10-09 08:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:00.186434012 +0000 UTC m=+1118.376962648" watchObservedRunningTime="2025-10-09 08:36:00.211491255 +0000 UTC m=+1118.402019891" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.212241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerStarted","Data":"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71"} Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.212322 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.215574 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zprz2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.217829 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.218405 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.218441 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.509101 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-66db9d8d8d-cz7s2"] Oct 09 08:36:00 crc kubenswrapper[4872]: E1009 08:36:00.509471 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" containerName="placement-db-sync" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.509483 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" containerName="placement-db-sync" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.509732 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" containerName="placement-db-sync" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.511113 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66db9d8d8d-cz7s2"] Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.511191 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.516313 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.516416 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.516738 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.516894 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-h4hr6" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.518733 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619723 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-public-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619769 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wd6d\" (UniqueName: \"kubernetes.io/projected/f4ad5d2f-42db-4d64-9d37-d37e89539c20-kube-api-access-5wd6d\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619849 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-internal-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619865 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-combined-ca-bundle\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619882 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-config-data\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619930 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-scripts\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.619957 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ad5d2f-42db-4d64-9d37-d37e89539c20-logs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.721354 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-internal-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.721437 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-combined-ca-bundle\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.721462 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-config-data\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.722581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-scripts\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.722625 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ad5d2f-42db-4d64-9d37-d37e89539c20-logs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.722728 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-public-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.722763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wd6d\" (UniqueName: \"kubernetes.io/projected/f4ad5d2f-42db-4d64-9d37-d37e89539c20-kube-api-access-5wd6d\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.724654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4ad5d2f-42db-4d64-9d37-d37e89539c20-logs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.732340 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-config-data\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.736243 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-scripts\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.740773 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-internal-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.741692 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-combined-ca-bundle\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.742073 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4ad5d2f-42db-4d64-9d37-d37e89539c20-public-tls-certs\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.752567 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wd6d\" (UniqueName: \"kubernetes.io/projected/f4ad5d2f-42db-4d64-9d37-d37e89539c20-kube-api-access-5wd6d\") pod \"placement-66db9d8d8d-cz7s2\" (UID: \"f4ad5d2f-42db-4d64-9d37-d37e89539c20\") " pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.875360 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:00 crc kubenswrapper[4872]: I1009 08:36:00.957832 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c68d5ff89-tcgwk"] Oct 09 08:36:00 crc kubenswrapper[4872]: W1009 08:36:00.973209 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7ec7531f_ab27_4417_a1f3_8b9a8f9fd7f0.slice/crio-6af1f0d0778edbae2de5c2726c5e312d0f943cf57e323de11fb69ea3efa54f55 WatchSource:0}: Error finding container 6af1f0d0778edbae2de5c2726c5e312d0f943cf57e323de11fb69ea3efa54f55: Status 404 returned error can't find the container with id 6af1f0d0778edbae2de5c2726c5e312d0f943cf57e323de11fb69ea3efa54f55 Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.255892 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8jxqn" event={"ID":"88cf1623-86ec-41eb-802f-c0996b7442be","Type":"ContainerStarted","Data":"0b4f7e667c969fd1c95e3e6f9770ccc34baf56a429957df46173bbf0b399d703"} Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.276539 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" event={"ID":"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c","Type":"ContainerStarted","Data":"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23"} Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.276760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.281244 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-8jxqn" podStartSLOduration=2.373692757 podStartE2EDuration="44.281228265s" podCreationTimestamp="2025-10-09 08:35:17 +0000 UTC" firstStartedPulling="2025-10-09 08:35:17.899175695 +0000 UTC m=+1076.089704321" lastFinishedPulling="2025-10-09 08:35:59.806711203 +0000 UTC m=+1117.997239829" observedRunningTime="2025-10-09 08:36:01.275414708 +0000 UTC m=+1119.465943334" watchObservedRunningTime="2025-10-09 08:36:01.281228265 +0000 UTC m=+1119.471756891" Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.304053 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerStarted","Data":"d258faa64da10e60d56fe7f96c08d75962ba9be75b1fff2a7b65f225fad9a31b"} Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.304093 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerStarted","Data":"95252d2cac742c3f41f8739ec31c6f49eee975bf4092730720e7f3e348b30f02"} Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.306555 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.312706 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c68d5ff89-tcgwk" event={"ID":"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0","Type":"ContainerStarted","Data":"6af1f0d0778edbae2de5c2726c5e312d0f943cf57e323de11fb69ea3efa54f55"} Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.340275 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" podStartSLOduration=10.340258389 podStartE2EDuration="10.340258389s" podCreationTimestamp="2025-10-09 08:35:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:01.319347616 +0000 UTC m=+1119.509876252" watchObservedRunningTime="2025-10-09 08:36:01.340258389 +0000 UTC m=+1119.530787015" Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.342517 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-66fb896578-k7hvx" podStartSLOduration=9.342504554 podStartE2EDuration="9.342504554s" podCreationTimestamp="2025-10-09 08:35:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:01.336621614 +0000 UTC m=+1119.527150240" watchObservedRunningTime="2025-10-09 08:36:01.342504554 +0000 UTC m=+1119.533033180" Oct 09 08:36:01 crc kubenswrapper[4872]: I1009 08:36:01.482151 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-66db9d8d8d-cz7s2"] Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.327145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c68d5ff89-tcgwk" event={"ID":"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0","Type":"ContainerStarted","Data":"3c4ec135a19cc0ad1275eb74d9c49700d9e400e076989c9cad1fbdade6c296c2"} Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.327585 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.327597 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c68d5ff89-tcgwk" event={"ID":"7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0","Type":"ContainerStarted","Data":"3947cfefccabadfcd55ab79b7537e05f31033d2d7f540885d2e95c8bc0554b60"} Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329562 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66db9d8d8d-cz7s2" event={"ID":"f4ad5d2f-42db-4d64-9d37-d37e89539c20","Type":"ContainerStarted","Data":"2aaf794044cbcf064465a76ba0a049bb8f46113ffcccdfb516b9c561947e4c6d"} Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329592 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66db9d8d8d-cz7s2" event={"ID":"f4ad5d2f-42db-4d64-9d37-d37e89539c20","Type":"ContainerStarted","Data":"d6f303fc7cbe6957c29db9ea3aa1aa3efc119acda734e556e7d9ffd21141285b"} Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329604 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329607 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329632 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.329615 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.350972 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c68d5ff89-tcgwk" podStartSLOduration=8.350947637 podStartE2EDuration="8.350947637s" podCreationTimestamp="2025-10-09 08:35:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:02.35003113 +0000 UTC m=+1120.540559776" watchObservedRunningTime="2025-10-09 08:36:02.350947637 +0000 UTC m=+1120.541476303" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.804405 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:02 crc kubenswrapper[4872]: I1009 08:36:02.807901 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:03 crc kubenswrapper[4872]: I1009 08:36:03.166518 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 08:36:03 crc kubenswrapper[4872]: I1009 08:36:03.352075 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-66db9d8d8d-cz7s2" event={"ID":"f4ad5d2f-42db-4d64-9d37-d37e89539c20","Type":"ContainerStarted","Data":"8b70317195ced04dd447dd5b1afe9822200bdeb353dbe7ba37af4e24acc78aa5"} Oct 09 08:36:03 crc kubenswrapper[4872]: I1009 08:36:03.352121 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:03 crc kubenswrapper[4872]: I1009 08:36:03.377998 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-66db9d8d8d-cz7s2" podStartSLOduration=3.377979476 podStartE2EDuration="3.377979476s" podCreationTimestamp="2025-10-09 08:36:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:03.376054451 +0000 UTC m=+1121.566583077" watchObservedRunningTime="2025-10-09 08:36:03.377979476 +0000 UTC m=+1121.568508092" Oct 09 08:36:03 crc kubenswrapper[4872]: I1009 08:36:03.987133 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 08:36:04 crc kubenswrapper[4872]: I1009 08:36:04.375479 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:04 crc kubenswrapper[4872]: I1009 08:36:04.375518 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:05 crc kubenswrapper[4872]: I1009 08:36:05.682872 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 09 08:36:05 crc kubenswrapper[4872]: I1009 08:36:05.769707 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5bf66bd496-2sdwc" podUID="6d37cb10-16dd-4360-aaf9-b90839973688" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.146:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.146:8443: connect: connection refused" Oct 09 08:36:06 crc kubenswrapper[4872]: I1009 08:36:06.392760 4872 generic.go:334] "Generic (PLEG): container finished" podID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" containerID="b2f2219e4942c73b94cb762a64e5ee5f127a4f829ea5c155feeb42e570e44b3e" exitCode=0 Oct 09 08:36:06 crc kubenswrapper[4872]: I1009 08:36:06.392856 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j52s6" event={"ID":"658ddbb1-1aaf-4e34-9a59-c6d01a87c505","Type":"ContainerDied","Data":"b2f2219e4942c73b94cb762a64e5ee5f127a4f829ea5c155feeb42e570e44b3e"} Oct 09 08:36:07 crc kubenswrapper[4872]: I1009 08:36:07.524845 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:36:07 crc kubenswrapper[4872]: I1009 08:36:07.593240 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:36:07 crc kubenswrapper[4872]: I1009 08:36:07.593513 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="dnsmasq-dns" containerID="cri-o://201aee04d487f4b7c89d007e33548199cf03d4ca4794849d8bc07b634a6c1d81" gracePeriod=10 Oct 09 08:36:08 crc kubenswrapper[4872]: I1009 08:36:08.158005 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:36:08 crc kubenswrapper[4872]: I1009 08:36:08.158289 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:36:08 crc kubenswrapper[4872]: I1009 08:36:08.411407 4872 generic.go:334] "Generic (PLEG): container finished" podID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerID="201aee04d487f4b7c89d007e33548199cf03d4ca4794849d8bc07b634a6c1d81" exitCode=0 Oct 09 08:36:08 crc kubenswrapper[4872]: I1009 08:36:08.411450 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" event={"ID":"7de46c6b-64ae-48c0-9485-5c2c8685e8b9","Type":"ContainerDied","Data":"201aee04d487f4b7c89d007e33548199cf03d4ca4794849d8bc07b634a6c1d81"} Oct 09 08:36:09 crc kubenswrapper[4872]: I1009 08:36:09.954342 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.148:5353: connect: connection refused" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.282553 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j52s6" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.421253 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data\") pod \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.421317 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle\") pod \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.421496 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs57t\" (UniqueName: \"kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t\") pod \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\" (UID: \"658ddbb1-1aaf-4e34-9a59-c6d01a87c505\") " Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.434932 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t" (OuterVolumeSpecName: "kube-api-access-xs57t") pod "658ddbb1-1aaf-4e34-9a59-c6d01a87c505" (UID: "658ddbb1-1aaf-4e34-9a59-c6d01a87c505"). InnerVolumeSpecName "kube-api-access-xs57t". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.436876 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "658ddbb1-1aaf-4e34-9a59-c6d01a87c505" (UID: "658ddbb1-1aaf-4e34-9a59-c6d01a87c505"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.460879 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-j52s6" event={"ID":"658ddbb1-1aaf-4e34-9a59-c6d01a87c505","Type":"ContainerDied","Data":"f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76"} Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.460923 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9983fcda1f88b9b752230a1d4763edcc33485bb0e6907b5158cf52109b4db76" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.460981 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-j52s6" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.504753 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "658ddbb1-1aaf-4e34-9a59-c6d01a87c505" (UID: "658ddbb1-1aaf-4e34-9a59-c6d01a87c505"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.523354 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.523389 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:10 crc kubenswrapper[4872]: I1009 08:36:10.523398 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs57t\" (UniqueName: \"kubernetes.io/projected/658ddbb1-1aaf-4e34-9a59-c6d01a87c505-kube-api-access-xs57t\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.473593 4872 generic.go:334] "Generic (PLEG): container finished" podID="88cf1623-86ec-41eb-802f-c0996b7442be" containerID="0b4f7e667c969fd1c95e3e6f9770ccc34baf56a429957df46173bbf0b399d703" exitCode=0 Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.473788 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8jxqn" event={"ID":"88cf1623-86ec-41eb-802f-c0996b7442be","Type":"ContainerDied","Data":"0b4f7e667c969fd1c95e3e6f9770ccc34baf56a429957df46173bbf0b399d703"} Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.589810 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7"] Oct 09 08:36:11 crc kubenswrapper[4872]: E1009 08:36:11.590325 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" containerName="barbican-db-sync" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.590352 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" containerName="barbican-db-sync" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.590631 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" containerName="barbican-db-sync" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.591863 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.599312 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.599312 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.599982 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-hml5z" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.614835 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-77bb59858c-hh8zx"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.616370 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.620364 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.646196 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77bb59858c-hh8zx"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.679686 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.687587 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.688962 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.720070 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748080 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2nzj\" (UniqueName: \"kubernetes.io/projected/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-kube-api-access-h2nzj\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data-custom\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748186 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-combined-ca-bundle\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748225 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data-custom\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748248 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-logs\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748280 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748326 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748361 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsdfr\" (UniqueName: \"kubernetes.io/projected/d6806fde-b186-4daa-b56d-e746704bf9a7-kube-api-access-lsdfr\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6806fde-b186-4daa-b56d-e746704bf9a7-logs\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.748423 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-combined-ca-bundle\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.766988 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.768473 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.772196 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.780152 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.850523 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.850581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2nzj\" (UniqueName: \"kubernetes.io/projected/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-kube-api-access-h2nzj\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.850660 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b49dm\" (UniqueName: \"kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.850692 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852783 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data-custom\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852824 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852850 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-combined-ca-bundle\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852924 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data-custom\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852950 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-logs\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.852988 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853056 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853085 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853123 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsdfr\" (UniqueName: \"kubernetes.io/projected/d6806fde-b186-4daa-b56d-e746704bf9a7-kube-api-access-lsdfr\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853165 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6806fde-b186-4daa-b56d-e746704bf9a7-logs\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-combined-ca-bundle\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853274 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.853706 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-logs\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.859554 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-combined-ca-bundle\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.859845 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6806fde-b186-4daa-b56d-e746704bf9a7-logs\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.864855 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-combined-ca-bundle\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.867618 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data-custom\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.872689 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2nzj\" (UniqueName: \"kubernetes.io/projected/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-kube-api-access-h2nzj\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.873360 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data-custom\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.884984 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsdfr\" (UniqueName: \"kubernetes.io/projected/d6806fde-b186-4daa-b56d-e746704bf9a7-kube-api-access-lsdfr\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.891382 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0126d82a-0297-4a4e-b75a-1e0a8f31d9d4-config-data\") pod \"barbican-worker-77bb59858c-hh8zx\" (UID: \"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4\") " pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.891565 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6806fde-b186-4daa-b56d-e746704bf9a7-config-data\") pod \"barbican-keystone-listener-5b8ccbfb5d-khqh7\" (UID: \"d6806fde-b186-4daa-b56d-e746704bf9a7\") " pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.928296 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.944669 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-77bb59858c-hh8zx" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956440 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j2qh\" (UniqueName: \"kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956549 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956601 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956627 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956678 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956724 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956746 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956812 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956860 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b49dm\" (UniqueName: \"kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956885 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.956909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.959021 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.959092 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.959238 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.961743 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.962178 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:11 crc kubenswrapper[4872]: I1009 08:36:11.981848 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b49dm\" (UniqueName: \"kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm\") pod \"dnsmasq-dns-688c87cc99-sbd8f\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.020557 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066022 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j2qh\" (UniqueName: \"kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066110 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066141 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066162 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066200 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.066808 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.072945 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.073120 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.074282 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.085819 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j2qh\" (UniqueName: \"kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh\") pod \"barbican-api-846c9bf8fd-2l4kt\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.098558 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.163576 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270013 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4dcs\" (UniqueName: \"kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270082 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270149 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270172 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270309 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.270356 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config\") pod \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\" (UID: \"7de46c6b-64ae-48c0-9485-5c2c8685e8b9\") " Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.283203 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs" (OuterVolumeSpecName: "kube-api-access-d4dcs") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "kube-api-access-d4dcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.336367 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.347454 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.382019 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.382119 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.382323 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.382337 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4dcs\" (UniqueName: \"kubernetes.io/projected/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-kube-api-access-d4dcs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.458197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config" (OuterVolumeSpecName: "config") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.501605 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.501934 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.520199 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.546209 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7de46c6b-64ae-48c0-9485-5c2c8685e8b9" (UID: "7de46c6b-64ae-48c0-9485-5c2c8685e8b9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.607170 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7de46c6b-64ae-48c0-9485-5c2c8685e8b9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.661260 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57c957c4ff-bvnfw" event={"ID":"7de46c6b-64ae-48c0-9485-5c2c8685e8b9","Type":"ContainerDied","Data":"ba4146edbfff7b3425a0d3b762411018e2baa2e5ec88ee2bdfe33c69a78b0460"} Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.661536 4872 scope.go:117] "RemoveContainer" containerID="201aee04d487f4b7c89d007e33548199cf03d4ca4794849d8bc07b634a6c1d81" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.701914 4872 scope.go:117] "RemoveContainer" containerID="547420e4687f983d560790af1d1488f17c5bde8571e7949ecedb24e7bb806350" Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.858361 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.869676 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57c957c4ff-bvnfw"] Oct 09 08:36:12 crc kubenswrapper[4872]: I1009 08:36:12.933286 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-77bb59858c-hh8zx"] Oct 09 08:36:12 crc kubenswrapper[4872]: W1009 08:36:12.949519 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0126d82a_0297_4a4e_b75a_1e0a8f31d9d4.slice/crio-4cabb46ae898c0b0f4233a751ce5e27e763df7b168a7248716c7e4963c4fceba WatchSource:0}: Error finding container 4cabb46ae898c0b0f4233a751ce5e27e763df7b168a7248716c7e4963c4fceba: Status 404 returned error can't find the container with id 4cabb46ae898c0b0f4233a751ce5e27e763df7b168a7248716c7e4963c4fceba Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.017400 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.088558 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.093924 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:13 crc kubenswrapper[4872]: W1009 08:36:13.101673 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c7f4d3a_a1da_4daf_a500_7f2e0a550975.slice/crio-9923e43e1df7466bb84b245cc69132fac0207058bf5532a17b4f966667d68ee6 WatchSource:0}: Error finding container 9923e43e1df7466bb84b245cc69132fac0207058bf5532a17b4f966667d68ee6: Status 404 returned error can't find the container with id 9923e43e1df7466bb84b245cc69132fac0207058bf5532a17b4f966667d68ee6 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.103052 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.116534 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.116733 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.116845 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.117329 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd598\" (UniqueName: \"kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.117411 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.117488 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.117655 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle\") pod \"88cf1623-86ec-41eb-802f-c0996b7442be\" (UID: \"88cf1623-86ec-41eb-802f-c0996b7442be\") " Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.118348 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/88cf1623-86ec-41eb-802f-c0996b7442be-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.133277 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598" (OuterVolumeSpecName: "kube-api-access-sd598") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "kube-api-access-sd598". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.133718 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts" (OuterVolumeSpecName: "scripts") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.138180 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.166847 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.220109 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.220134 4872 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.220145 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd598\" (UniqueName: \"kubernetes.io/projected/88cf1623-86ec-41eb-802f-c0996b7442be-kube-api-access-sd598\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.220155 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.220234 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data" (OuterVolumeSpecName: "config-data") pod "88cf1623-86ec-41eb-802f-c0996b7442be" (UID: "88cf1623-86ec-41eb-802f-c0996b7442be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.323360 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/88cf1623-86ec-41eb-802f-c0996b7442be-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.540435 4872 generic.go:334] "Generic (PLEG): container finished" podID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerID="5839dc14d07fab733ae706a2c3966f50eaef9b7e2279569489a735d6491fb1cc" exitCode=0 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.540515 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" event={"ID":"2c7f4d3a-a1da-4daf-a500-7f2e0a550975","Type":"ContainerDied","Data":"5839dc14d07fab733ae706a2c3966f50eaef9b7e2279569489a735d6491fb1cc"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.540544 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" event={"ID":"2c7f4d3a-a1da-4daf-a500-7f2e0a550975","Type":"ContainerStarted","Data":"9923e43e1df7466bb84b245cc69132fac0207058bf5532a17b4f966667d68ee6"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.550839 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerStarted","Data":"471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.550879 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerStarted","Data":"a32606e1cf9b49455fdacb1ad04f68c74d0ca1c77436f1e12165f4bcf0f46b8b"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.554200 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-8jxqn" event={"ID":"88cf1623-86ec-41eb-802f-c0996b7442be","Type":"ContainerDied","Data":"87a95a7545445b96d64fd264e33d71f8ed355bff740eab2a796877f4675bc0f3"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.554232 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87a95a7545445b96d64fd264e33d71f8ed355bff740eab2a796877f4675bc0f3" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.554300 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-8jxqn" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.557933 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" event={"ID":"d6806fde-b186-4daa-b56d-e746704bf9a7","Type":"ContainerStarted","Data":"c3a128dcb5382eca69c20473405ca2ccf1546f5fe4299675c7b475408a69adcf"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.574183 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bb59858c-hh8zx" event={"ID":"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4","Type":"ContainerStarted","Data":"4cabb46ae898c0b0f4233a751ce5e27e763df7b168a7248716c7e4963c4fceba"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.582108 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerStarted","Data":"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50"} Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.582689 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-central-agent" containerID="cri-o://10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b" gracePeriod=30 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.582936 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.583005 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="proxy-httpd" containerID="cri-o://301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50" gracePeriod=30 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.583019 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="sg-core" containerID="cri-o://2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71" gracePeriod=30 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.583168 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-notification-agent" containerID="cri-o://c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e" gracePeriod=30 Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.608210 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.976325111 podStartE2EDuration="57.608184659s" podCreationTimestamp="2025-10-09 08:35:16 +0000 UTC" firstStartedPulling="2025-10-09 08:35:17.634998461 +0000 UTC m=+1075.825527087" lastFinishedPulling="2025-10-09 08:36:12.266858009 +0000 UTC m=+1130.457386635" observedRunningTime="2025-10-09 08:36:13.604824652 +0000 UTC m=+1131.795353288" watchObservedRunningTime="2025-10-09 08:36:13.608184659 +0000 UTC m=+1131.798713305" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.723992 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:13 crc kubenswrapper[4872]: E1009 08:36:13.724343 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="dnsmasq-dns" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.724359 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="dnsmasq-dns" Oct 09 08:36:13 crc kubenswrapper[4872]: E1009 08:36:13.724366 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" containerName="cinder-db-sync" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.724372 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" containerName="cinder-db-sync" Oct 09 08:36:13 crc kubenswrapper[4872]: E1009 08:36:13.724399 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="init" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.724405 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="init" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.724567 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" containerName="dnsmasq-dns" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.724583 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" containerName="cinder-db-sync" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.725442 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.734385 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.734846 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.735313 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.735461 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-kwdcb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.758922 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.769082 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.806509 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.808002 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851054 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851115 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851170 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg2wh\" (UniqueName: \"kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851195 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851223 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.851252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.853496 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.940116 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.941661 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.943196 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.952775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.952856 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.952918 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.952950 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.953055 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.954968 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955012 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955094 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955141 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pchsv\" (UniqueName: \"kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955169 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955260 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.955404 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wg2wh\" (UniqueName: \"kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.956220 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.958846 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.960045 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.962115 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.965392 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.972409 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:13 crc kubenswrapper[4872]: I1009 08:36:13.981892 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wg2wh\" (UniqueName: \"kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh\") pod \"cinder-scheduler-0\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.056939 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pchsv\" (UniqueName: \"kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057387 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057499 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057614 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gwdd\" (UniqueName: \"kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057709 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057784 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057857 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.057943 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.058047 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.058259 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.058340 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.058404 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.058783 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.059249 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.059776 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.063672 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.065177 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.068246 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.081366 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pchsv\" (UniqueName: \"kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv\") pod \"dnsmasq-dns-6bb4fc677f-glltb\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.132095 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.160671 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161269 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161353 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161510 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161666 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gwdd\" (UniqueName: \"kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161700 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161751 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.161873 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.162166 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.167094 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.170379 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.175486 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.217604 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.224194 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gwdd\" (UniqueName: \"kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd\") pod \"cinder-api-0\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.382324 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.475512 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de46c6b-64ae-48c0-9485-5c2c8685e8b9" path="/var/lib/kubelet/pods/7de46c6b-64ae-48c0-9485-5c2c8685e8b9/volumes" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.574102 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.631942 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerID="301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50" exitCode=0 Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.632008 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerID="2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71" exitCode=2 Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.632019 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerID="10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b" exitCode=0 Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.632036 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerDied","Data":"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50"} Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.632075 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerDied","Data":"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71"} Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.632088 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerDied","Data":"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b"} Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.647671 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" event={"ID":"2c7f4d3a-a1da-4daf-a500-7f2e0a550975","Type":"ContainerStarted","Data":"3eb946f9499fa7e30080dc09cbf0d2c861977eabb4642e1e7b037e72f0de402b"} Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.647755 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="dnsmasq-dns" containerID="cri-o://3eb946f9499fa7e30080dc09cbf0d2c861977eabb4642e1e7b037e72f0de402b" gracePeriod=10 Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.648295 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.657013 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerStarted","Data":"197aaf63fc4c4afddb768ee10ed2e17d1d4983b192ac4fc05912e4cd5b72ef9b"} Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.657470 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.657483 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.666140 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" podStartSLOduration=3.6661255600000002 podStartE2EDuration="3.66612556s" podCreationTimestamp="2025-10-09 08:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:14.6657602 +0000 UTC m=+1132.856288826" watchObservedRunningTime="2025-10-09 08:36:14.66612556 +0000 UTC m=+1132.856654186" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.708300 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-846c9bf8fd-2l4kt" podStartSLOduration=3.708260276 podStartE2EDuration="3.708260276s" podCreationTimestamp="2025-10-09 08:36:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:14.683527232 +0000 UTC m=+1132.874055858" watchObservedRunningTime="2025-10-09 08:36:14.708260276 +0000 UTC m=+1132.898788912" Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.785119 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:36:14 crc kubenswrapper[4872]: I1009 08:36:14.916518 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:15 crc kubenswrapper[4872]: W1009 08:36:15.613981 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fcf1521_b4c8_4a2a_be8f_31c48c2960ca.slice/crio-4453d37f68179950f835eb0a781feacfa69dd801f7e98e9fcf5e7943b3cb8b9f WatchSource:0}: Error finding container 4453d37f68179950f835eb0a781feacfa69dd801f7e98e9fcf5e7943b3cb8b9f: Status 404 returned error can't find the container with id 4453d37f68179950f835eb0a781feacfa69dd801f7e98e9fcf5e7943b3cb8b9f Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.668881 4872 generic.go:334] "Generic (PLEG): container finished" podID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerID="fa7e50b7bf9d20e414407cefeec9c0a3953becbf8defe0943a0370c0802cdafb" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.668907 4872 generic.go:334] "Generic (PLEG): container finished" podID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerID="401c2b173285cfab1735af4da1a043607f76c7e975f58eb132c70a3a0b12beaf" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.668962 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerDied","Data":"fa7e50b7bf9d20e414407cefeec9c0a3953becbf8defe0943a0370c0802cdafb"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.669009 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerDied","Data":"401c2b173285cfab1735af4da1a043607f76c7e975f58eb132c70a3a0b12beaf"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.670243 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerStarted","Data":"4453d37f68179950f835eb0a781feacfa69dd801f7e98e9fcf5e7943b3cb8b9f"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.671628 4872 generic.go:334] "Generic (PLEG): container finished" podID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerID="3eb946f9499fa7e30080dc09cbf0d2c861977eabb4642e1e7b037e72f0de402b" exitCode=0 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.671657 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" event={"ID":"2c7f4d3a-a1da-4daf-a500-7f2e0a550975","Type":"ContainerDied","Data":"3eb946f9499fa7e30080dc09cbf0d2c861977eabb4642e1e7b037e72f0de402b"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.673218 4872 generic.go:334] "Generic (PLEG): container finished" podID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerID="bbd62b7f0c0a68e436020bd4fc66acbecd5e2cea57db6a55cdafe0ae317be93e" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.673240 4872 generic.go:334] "Generic (PLEG): container finished" podID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerID="0e540636c822a0b88202d76d69fbb28add2b27d431302b713682c776fe3fdb14" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.673285 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerDied","Data":"bbd62b7f0c0a68e436020bd4fc66acbecd5e2cea57db6a55cdafe0ae317be93e"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.673304 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerDied","Data":"0e540636c822a0b88202d76d69fbb28add2b27d431302b713682c776fe3fdb14"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.678474 4872 generic.go:334] "Generic (PLEG): container finished" podID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerID="db797175d5e966c6e067219ff4a95c22c14a5f3bd5b25bc32545c9ea9c11554c" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.678497 4872 generic.go:334] "Generic (PLEG): container finished" podID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerID="672a9e02184b99f9a968b3657325a33f4757bc40a20b94f57965e7845b8d7a13" exitCode=137 Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.678757 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerDied","Data":"db797175d5e966c6e067219ff4a95c22c14a5f3bd5b25bc32545c9ea9c11554c"} Oct 09 08:36:15 crc kubenswrapper[4872]: I1009 08:36:15.678804 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerDied","Data":"672a9e02184b99f9a968b3657325a33f4757bc40a20b94f57965e7845b8d7a13"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.212435 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.311996 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.312035 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b49dm\" (UniqueName: \"kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.312060 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.312089 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.312155 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.312215 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb\") pod \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\" (UID: \"2c7f4d3a-a1da-4daf-a500-7f2e0a550975\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.356842 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm" (OuterVolumeSpecName: "kube-api-access-b49dm") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "kube-api-access-b49dm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.414807 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b49dm\" (UniqueName: \"kubernetes.io/projected/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-kube-api-access-b49dm\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.592738 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.668419 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.685151 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.697462 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" event={"ID":"d6806fde-b186-4daa-b56d-e746704bf9a7","Type":"ContainerStarted","Data":"a34081bb34fffb98dd00f317802c61d91df6707377a310630a1274f276544dde"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.702278 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66cb5748f-phdrk" event={"ID":"bb11c541-47a1-4494-9d1d-b8d7272c376f","Type":"ContainerDied","Data":"d9f858e3fdcc797ab275aff901416be678fd2205f6be672d53d8cd55d0af019a"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.702310 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f858e3fdcc797ab275aff901416be678fd2205f6be672d53d8cd55d0af019a" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.704565 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerStarted","Data":"e20bd1006477608efcb7b91bd3cf777cb95ea68a5f7e43665d742b83925cd44f"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.704608 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerStarted","Data":"278f3f0426596fa38aaa36894a5b9d7c2b9780f778fe1431c0fb812e166bebec"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.706800 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bb59858c-hh8zx" event={"ID":"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4","Type":"ContainerStarted","Data":"1c570cbbd1b5a52303dbeb07d84ed052bb285fd1a45a9b8bad3a5afaf627bd93"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.710331 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" event={"ID":"2c7f4d3a-a1da-4daf-a500-7f2e0a550975","Type":"ContainerDied","Data":"9923e43e1df7466bb84b245cc69132fac0207058bf5532a17b4f966667d68ee6"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.710539 4872 scope.go:117] "RemoveContainer" containerID="3eb946f9499fa7e30080dc09cbf0d2c861977eabb4642e1e7b037e72f0de402b" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.710501 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688c87cc99-sbd8f" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.711508 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerStarted","Data":"4a9ef58e4c917b6f04ea8001269d595df872585c60b0df59cae210b6e949987f"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.718922 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-648775cb6f-77vfj" event={"ID":"db4e89a0-ad1c-4a35-9f6d-5d13de37d742","Type":"ContainerDied","Data":"59272a8444a79613e701d551f071a192fda2e2b4e9981c9939b3a7f8c4a446d0"} Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.719214 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-648775cb6f-77vfj" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.720664 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.727253 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs\") pod \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.727333 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts\") pod \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.727421 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data\") pod \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.727495 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key\") pod \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.727622 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd\") pod \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\" (UID: \"db4e89a0-ad1c-4a35-9f6d-5d13de37d742\") " Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.728113 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.728127 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.728136 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:16 crc kubenswrapper[4872]: I1009 08:36:16.729874 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs" (OuterVolumeSpecName: "logs") pod "db4e89a0-ad1c-4a35-9f6d-5d13de37d742" (UID: "db4e89a0-ad1c-4a35-9f6d-5d13de37d742"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.732889 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config" (OuterVolumeSpecName: "config") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.732890 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd" (OuterVolumeSpecName: "kube-api-access-cqvkd") pod "db4e89a0-ad1c-4a35-9f6d-5d13de37d742" (UID: "db4e89a0-ad1c-4a35-9f6d-5d13de37d742"). InnerVolumeSpecName "kube-api-access-cqvkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.734343 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "db4e89a0-ad1c-4a35-9f6d-5d13de37d742" (UID: "db4e89a0-ad1c-4a35-9f6d-5d13de37d742"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.746740 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.747103 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.748579 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2c7f4d3a-a1da-4daf-a500-7f2e0a550975" (UID: "2c7f4d3a-a1da-4daf-a500-7f2e0a550975"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.753571 4872 scope.go:117] "RemoveContainer" containerID="5839dc14d07fab733ae706a2c3966f50eaef9b7e2279569489a735d6491fb1cc" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.760511 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts" (OuterVolumeSpecName: "scripts") pod "db4e89a0-ad1c-4a35-9f6d-5d13de37d742" (UID: "db4e89a0-ad1c-4a35-9f6d-5d13de37d742"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.773120 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data" (OuterVolumeSpecName: "config-data") pod "db4e89a0-ad1c-4a35-9f6d-5d13de37d742" (UID: "db4e89a0-ad1c-4a35-9f6d-5d13de37d742"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.808192 4872 scope.go:117] "RemoveContainer" containerID="db797175d5e966c6e067219ff4a95c22c14a5f3bd5b25bc32545c9ea9c11554c" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829051 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs\") pod \"d40652a2-f7f1-429c-9f66-061b4c8350f7\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829161 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n89n\" (UniqueName: \"kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n\") pod \"bb11c541-47a1-4494-9d1d-b8d7272c376f\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829190 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts\") pod \"d40652a2-f7f1-429c-9f66-061b4c8350f7\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829243 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key\") pod \"d40652a2-f7f1-429c-9f66-061b4c8350f7\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829299 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts\") pod \"bb11c541-47a1-4494-9d1d-b8d7272c376f\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829371 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key\") pod \"bb11c541-47a1-4494-9d1d-b8d7272c376f\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829401 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data\") pod \"d40652a2-f7f1-429c-9f66-061b4c8350f7\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829444 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data\") pod \"bb11c541-47a1-4494-9d1d-b8d7272c376f\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbljx\" (UniqueName: \"kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx\") pod \"d40652a2-f7f1-429c-9f66-061b4c8350f7\" (UID: \"d40652a2-f7f1-429c-9f66-061b4c8350f7\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.829514 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs\") pod \"bb11c541-47a1-4494-9d1d-b8d7272c376f\" (UID: \"bb11c541-47a1-4494-9d1d-b8d7272c376f\") " Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830102 4872 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830116 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830126 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-kube-api-access-cqvkd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830137 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830147 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830155 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c7f4d3a-a1da-4daf-a500-7f2e0a550975-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.830164 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db4e89a0-ad1c-4a35-9f6d-5d13de37d742-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.831103 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs" (OuterVolumeSpecName: "logs") pod "bb11c541-47a1-4494-9d1d-b8d7272c376f" (UID: "bb11c541-47a1-4494-9d1d-b8d7272c376f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.831140 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs" (OuterVolumeSpecName: "logs") pod "d40652a2-f7f1-429c-9f66-061b4c8350f7" (UID: "d40652a2-f7f1-429c-9f66-061b4c8350f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.833959 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n" (OuterVolumeSpecName: "kube-api-access-6n89n") pod "bb11c541-47a1-4494-9d1d-b8d7272c376f" (UID: "bb11c541-47a1-4494-9d1d-b8d7272c376f"). InnerVolumeSpecName "kube-api-access-6n89n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.837578 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "d40652a2-f7f1-429c-9f66-061b4c8350f7" (UID: "d40652a2-f7f1-429c-9f66-061b4c8350f7"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.851572 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "bb11c541-47a1-4494-9d1d-b8d7272c376f" (UID: "bb11c541-47a1-4494-9d1d-b8d7272c376f"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.856171 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx" (OuterVolumeSpecName: "kube-api-access-zbljx") pod "d40652a2-f7f1-429c-9f66-061b4c8350f7" (UID: "d40652a2-f7f1-429c-9f66-061b4c8350f7"). InnerVolumeSpecName "kube-api-access-zbljx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.874101 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data" (OuterVolumeSpecName: "config-data") pod "bb11c541-47a1-4494-9d1d-b8d7272c376f" (UID: "bb11c541-47a1-4494-9d1d-b8d7272c376f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.875098 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts" (OuterVolumeSpecName: "scripts") pod "bb11c541-47a1-4494-9d1d-b8d7272c376f" (UID: "bb11c541-47a1-4494-9d1d-b8d7272c376f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.875944 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data" (OuterVolumeSpecName: "config-data") pod "d40652a2-f7f1-429c-9f66-061b4c8350f7" (UID: "d40652a2-f7f1-429c-9f66-061b4c8350f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.896874 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts" (OuterVolumeSpecName: "scripts") pod "d40652a2-f7f1-429c-9f66-061b4c8350f7" (UID: "d40652a2-f7f1-429c-9f66-061b4c8350f7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.931978 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932005 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbljx\" (UniqueName: \"kubernetes.io/projected/d40652a2-f7f1-429c-9f66-061b4c8350f7-kube-api-access-zbljx\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932017 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb11c541-47a1-4494-9d1d-b8d7272c376f-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932026 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d40652a2-f7f1-429c-9f66-061b4c8350f7-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932034 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n89n\" (UniqueName: \"kubernetes.io/projected/bb11c541-47a1-4494-9d1d-b8d7272c376f-kube-api-access-6n89n\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932043 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932051 4872 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/d40652a2-f7f1-429c-9f66-061b4c8350f7-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932059 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bb11c541-47a1-4494-9d1d-b8d7272c376f-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932066 4872 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bb11c541-47a1-4494-9d1d-b8d7272c376f-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:16.932074 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d40652a2-f7f1-429c-9f66-061b4c8350f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.006273 4872 scope.go:117] "RemoveContainer" containerID="672a9e02184b99f9a968b3657325a33f4757bc40a20b94f57965e7845b8d7a13" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.237207 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.245804 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688c87cc99-sbd8f"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.257908 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.275826 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-648775cb6f-77vfj"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.731607 4872 generic.go:334] "Generic (PLEG): container finished" podID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerID="e20bd1006477608efcb7b91bd3cf777cb95ea68a5f7e43665d742b83925cd44f" exitCode=0 Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.732692 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerDied","Data":"e20bd1006477608efcb7b91bd3cf777cb95ea68a5f7e43665d742b83925cd44f"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.732757 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.732774 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerStarted","Data":"dfe5352f1ca2fd7daae4cceb44cada175049f883c7a926b8c9de8f52508e12c6"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.736576 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-77bb59858c-hh8zx" event={"ID":"0126d82a-0297-4a4e-b75a-1e0a8f31d9d4","Type":"ContainerStarted","Data":"4085b8ad1d13dc9e3db8f7630ebb45eda30a44bb6bc15bf8550d6476944ba26e"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.739301 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerStarted","Data":"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.743459 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5749cd8d8c-b6qs4" event={"ID":"d40652a2-f7f1-429c-9f66-061b4c8350f7","Type":"ContainerDied","Data":"1efa65c66da0c21a141f5ac90fbf7670440bf7b5b5b8593409e9b20a33caa83f"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.743513 4872 scope.go:117] "RemoveContainer" containerID="bbd62b7f0c0a68e436020bd4fc66acbecd5e2cea57db6a55cdafe0ae317be93e" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.743696 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5749cd8d8c-b6qs4" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.749386 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66cb5748f-phdrk" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.750044 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" event={"ID":"d6806fde-b186-4daa-b56d-e746704bf9a7","Type":"ContainerStarted","Data":"c5eda0b61c9a2c8c5a00594b759005c36b77f404112b6384414eceead0803cd7"} Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.770751 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" podStartSLOduration=4.770726595 podStartE2EDuration="4.770726595s" podCreationTimestamp="2025-10-09 08:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:17.758207544 +0000 UTC m=+1135.948736170" watchObservedRunningTime="2025-10-09 08:36:17.770726595 +0000 UTC m=+1135.961255231" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.792249 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-77bb59858c-hh8zx" podStartSLOduration=3.8591340499999998 podStartE2EDuration="6.792228026s" podCreationTimestamp="2025-10-09 08:36:11 +0000 UTC" firstStartedPulling="2025-10-09 08:36:12.953989939 +0000 UTC m=+1131.144518565" lastFinishedPulling="2025-10-09 08:36:15.887083915 +0000 UTC m=+1134.077612541" observedRunningTime="2025-10-09 08:36:17.782714441 +0000 UTC m=+1135.973243077" watchObservedRunningTime="2025-10-09 08:36:17.792228026 +0000 UTC m=+1135.982756652" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.826059 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5b8ccbfb5d-khqh7" podStartSLOduration=3.949660203 podStartE2EDuration="6.826040362s" podCreationTimestamp="2025-10-09 08:36:11 +0000 UTC" firstStartedPulling="2025-10-09 08:36:13.010720617 +0000 UTC m=+1131.201249243" lastFinishedPulling="2025-10-09 08:36:15.887100776 +0000 UTC m=+1134.077629402" observedRunningTime="2025-10-09 08:36:17.822448368 +0000 UTC m=+1136.012976994" watchObservedRunningTime="2025-10-09 08:36:17.826040362 +0000 UTC m=+1136.016568988" Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.888773 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.903007 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-5749cd8d8c-b6qs4"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.914710 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.922220 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66cb5748f-phdrk"] Oct 09 08:36:17 crc kubenswrapper[4872]: I1009 08:36:17.941141 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.011527 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.018605 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.028746 4872 scope.go:117] "RemoveContainer" containerID="0e540636c822a0b88202d76d69fbb28add2b27d431302b713682c776fe3fdb14" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.523833 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" path="/var/lib/kubelet/pods/2c7f4d3a-a1da-4daf-a500-7f2e0a550975/volumes" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.526339 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" path="/var/lib/kubelet/pods/bb11c541-47a1-4494-9d1d-b8d7272c376f/volumes" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.527255 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" path="/var/lib/kubelet/pods/d40652a2-f7f1-429c-9f66-061b4c8350f7/volumes" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.528070 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" path="/var/lib/kubelet/pods/db4e89a0-ad1c-4a35-9f6d-5d13de37d742/volumes" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.745791 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58c4bcb5d6-tl5n8"] Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746200 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746216 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746226 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="dnsmasq-dns" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746236 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="dnsmasq-dns" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746250 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746257 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746277 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746284 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746301 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746308 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746326 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746334 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746347 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746356 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: E1009 08:36:18.746378 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="init" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746386 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="init" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746581 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c7f4d3a-a1da-4daf-a500-7f2e0a550975" containerName="dnsmasq-dns" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746597 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746614 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746624 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb11c541-47a1-4494-9d1d-b8d7272c376f" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746635 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746677 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d40652a2-f7f1-429c-9f66-061b4c8350f7" containerName="horizon" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.746688 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="db4e89a0-ad1c-4a35-9f6d-5d13de37d742" containerName="horizon-log" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.754412 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.756919 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.757130 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.802003 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerStarted","Data":"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e"} Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.803430 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58c4bcb5d6-tl5n8"] Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.817441 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerStarted","Data":"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a"} Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.818161 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.854936 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.854916124 podStartE2EDuration="5.854916124s" podCreationTimestamp="2025-10-09 08:36:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:18.851735442 +0000 UTC m=+1137.042264088" watchObservedRunningTime="2025-10-09 08:36:18.854916124 +0000 UTC m=+1137.045444750" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904538 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-internal-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data-custom\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904686 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zsrj\" (UniqueName: \"kubernetes.io/projected/4fa33440-7187-4d50-8fe8-b7d6af5db927-kube-api-access-4zsrj\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904740 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-combined-ca-bundle\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904919 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-public-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904966 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fa33440-7187-4d50-8fe8-b7d6af5db927-logs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:18 crc kubenswrapper[4872]: I1009 08:36:18.904989 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006010 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-internal-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006069 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data-custom\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006104 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zsrj\" (UniqueName: \"kubernetes.io/projected/4fa33440-7187-4d50-8fe8-b7d6af5db927-kube-api-access-4zsrj\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006136 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-combined-ca-bundle\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006219 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-public-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fa33440-7187-4d50-8fe8-b7d6af5db927-logs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006818 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4fa33440-7187-4d50-8fe8-b7d6af5db927-logs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.006920 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.011631 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-combined-ca-bundle\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.011790 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-internal-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.013184 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-public-tls-certs\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.014081 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data-custom\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.015061 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa33440-7187-4d50-8fe8-b7d6af5db927-config-data\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.028261 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zsrj\" (UniqueName: \"kubernetes.io/projected/4fa33440-7187-4d50-8fe8-b7d6af5db927-kube-api-access-4zsrj\") pod \"barbican-api-58c4bcb5d6-tl5n8\" (UID: \"4fa33440-7187-4d50-8fe8-b7d6af5db927\") " pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.077709 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.693961 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.713088 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58c4bcb5d6-tl5n8"] Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.832978 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833040 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833068 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k2sj8\" (UniqueName: \"kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833146 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833224 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833310 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833389 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml\") pod \"a4ba9d6d-2798-4832-9260-e055e35818b1\" (UID: \"a4ba9d6d-2798-4832-9260-e055e35818b1\") " Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.833551 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.834029 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.834055 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.838344 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" event={"ID":"4fa33440-7187-4d50-8fe8-b7d6af5db927","Type":"ContainerStarted","Data":"86b0acc688f65e6f5681a8a7b5c1c483ffcaaca9844759ebee94594d1e7e89cb"} Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.838333 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts" (OuterVolumeSpecName: "scripts") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.843821 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8" (OuterVolumeSpecName: "kube-api-access-k2sj8") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "kube-api-access-k2sj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.844331 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerStarted","Data":"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78"} Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.857609 4872 generic.go:334] "Generic (PLEG): container finished" podID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerID="c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e" exitCode=0 Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.857863 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api-log" containerID="cri-o://7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" gracePeriod=30 Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.857903 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerDied","Data":"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e"} Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.857983 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a4ba9d6d-2798-4832-9260-e055e35818b1","Type":"ContainerDied","Data":"b8851cdb126658499d04225e64dc6643a88d34f356b3e27164f8ec6b80ba7580"} Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.858024 4872 scope.go:117] "RemoveContainer" containerID="301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.858081 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api" containerID="cri-o://034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" gracePeriod=30 Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.858118 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.879278 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.732071117 podStartE2EDuration="6.874629281s" podCreationTimestamp="2025-10-09 08:36:13 +0000 UTC" firstStartedPulling="2025-10-09 08:36:15.615500187 +0000 UTC m=+1133.806028823" lastFinishedPulling="2025-10-09 08:36:16.758058361 +0000 UTC m=+1134.948586987" observedRunningTime="2025-10-09 08:36:19.868547656 +0000 UTC m=+1138.059076282" watchObservedRunningTime="2025-10-09 08:36:19.874629281 +0000 UTC m=+1138.065157917" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.907226 4872 scope.go:117] "RemoveContainer" containerID="2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.918749 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.931996 4872 scope.go:117] "RemoveContainer" containerID="c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.936428 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.936453 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a4ba9d6d-2798-4832-9260-e055e35818b1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.936462 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.936470 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k2sj8\" (UniqueName: \"kubernetes.io/projected/a4ba9d6d-2798-4832-9260-e055e35818b1-kube-api-access-k2sj8\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.963284 4872 scope.go:117] "RemoveContainer" containerID="10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b" Oct 09 08:36:19 crc kubenswrapper[4872]: I1009 08:36:19.979834 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.020865 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data" (OuterVolumeSpecName: "config-data") pod "a4ba9d6d-2798-4832-9260-e055e35818b1" (UID: "a4ba9d6d-2798-4832-9260-e055e35818b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.024922 4872 scope.go:117] "RemoveContainer" containerID="301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.025629 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50\": container with ID starting with 301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50 not found: ID does not exist" containerID="301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.025686 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50"} err="failed to get container status \"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50\": rpc error: code = NotFound desc = could not find container \"301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50\": container with ID starting with 301be26ded5865d3f80a753fb78fc8978f5d7158456be03db09681c3108afc50 not found: ID does not exist" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.025715 4872 scope.go:117] "RemoveContainer" containerID="2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.027079 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71\": container with ID starting with 2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71 not found: ID does not exist" containerID="2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.027115 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71"} err="failed to get container status \"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71\": rpc error: code = NotFound desc = could not find container \"2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71\": container with ID starting with 2c436c6f22f771da60e62b12246089cbb8486fe2470a06a2b8dde225ed3b8c71 not found: ID does not exist" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.027142 4872 scope.go:117] "RemoveContainer" containerID="c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.027480 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e\": container with ID starting with c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e not found: ID does not exist" containerID="c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.027511 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e"} err="failed to get container status \"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e\": rpc error: code = NotFound desc = could not find container \"c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e\": container with ID starting with c1b71272255ba4e1b33280eb34cbde6ed3be21e7f98aa0289321eb5172d9f39e not found: ID does not exist" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.027530 4872 scope.go:117] "RemoveContainer" containerID="10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.027905 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b\": container with ID starting with 10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b not found: ID does not exist" containerID="10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.027934 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b"} err="failed to get container status \"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b\": rpc error: code = NotFound desc = could not find container \"10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b\": container with ID starting with 10c8e148411d6dc2b472efa0b241d59087eaa475c896a00f6ccf631e2581271b not found: ID does not exist" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.045410 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.045797 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a4ba9d6d-2798-4832-9260-e055e35818b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.343036 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.355094 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372158 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.372541 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="proxy-httpd" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372558 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="proxy-httpd" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.372569 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-central-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372575 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-central-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.372588 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="sg-core" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372595 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="sg-core" Oct 09 08:36:20 crc kubenswrapper[4872]: E1009 08:36:20.372605 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-notification-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372611 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-notification-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372840 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="proxy-httpd" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372855 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-notification-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372875 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="sg-core" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.372888 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" containerName="ceilometer-central-agent" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.374476 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.382196 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.382384 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.400949 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455039 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455073 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455147 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fz66z\" (UniqueName: \"kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455178 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455197 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455215 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.455228 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.477107 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4ba9d6d-2798-4832-9260-e055e35818b1" path="/var/lib/kubelet/pods/a4ba9d6d-2798-4832-9260-e055e35818b1/volumes" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.556867 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557208 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557424 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fz66z\" (UniqueName: \"kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557533 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557751 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557905 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.557995 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.560703 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.568781 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.569540 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.572122 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.585584 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.602275 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fz66z\" (UniqueName: \"kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.603369 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data\") pod \"ceilometer-0\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.712804 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.743249 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.872326 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.872840 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.872963 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.872994 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.873084 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gwdd\" (UniqueName: \"kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.873111 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.875212 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data\") pod \"e4866439-dbe8-4715-afda-5965d455af54\" (UID: \"e4866439-dbe8-4715-afda-5965d455af54\") " Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.878500 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs" (OuterVolumeSpecName: "logs") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.878811 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.886128 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts" (OuterVolumeSpecName: "scripts") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.887053 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" event={"ID":"4fa33440-7187-4d50-8fe8-b7d6af5db927","Type":"ContainerStarted","Data":"d5efad62ef99aee3f472d72ab6eeb21db3a63790ff444646fd97e9035111ab6e"} Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.887095 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" event={"ID":"4fa33440-7187-4d50-8fe8-b7d6af5db927","Type":"ContainerStarted","Data":"86d802c5d06c3ffaa8b28e71555c97a3c783fc2477fd02ce6f0d5eb39674959b"} Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.890604 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.906668 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd" (OuterVolumeSpecName: "kube-api-access-5gwdd") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "kube-api-access-5gwdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.919196 4872 generic.go:334] "Generic (PLEG): container finished" podID="e4866439-dbe8-4715-afda-5965d455af54" containerID="034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" exitCode=0 Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.919276 4872 generic.go:334] "Generic (PLEG): container finished" podID="e4866439-dbe8-4715-afda-5965d455af54" containerID="7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" exitCode=143 Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.920690 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.921314 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerDied","Data":"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a"} Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.921348 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerDied","Data":"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583"} Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.921363 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4866439-dbe8-4715-afda-5965d455af54","Type":"ContainerDied","Data":"4a9ef58e4c917b6f04ea8001269d595df872585c60b0df59cae210b6e949987f"} Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.921383 4872 scope.go:117] "RemoveContainer" containerID="034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.978762 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gwdd\" (UniqueName: \"kubernetes.io/projected/e4866439-dbe8-4715-afda-5965d455af54-kube-api-access-5gwdd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.978790 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.978801 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4866439-dbe8-4715-afda-5965d455af54-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.978809 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:20 crc kubenswrapper[4872]: I1009 08:36:20.978817 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4866439-dbe8-4715-afda-5965d455af54-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.006511 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.012306 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.020916 4872 scope.go:117] "RemoveContainer" containerID="7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.039828 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data" (OuterVolumeSpecName: "config-data") pod "e4866439-dbe8-4715-afda-5965d455af54" (UID: "e4866439-dbe8-4715-afda-5965d455af54"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.062842 4872 scope.go:117] "RemoveContainer" containerID="034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" Oct 09 08:36:21 crc kubenswrapper[4872]: E1009 08:36:21.066956 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a\": container with ID starting with 034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a not found: ID does not exist" containerID="034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.067008 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a"} err="failed to get container status \"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a\": rpc error: code = NotFound desc = could not find container \"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a\": container with ID starting with 034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a not found: ID does not exist" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.067040 4872 scope.go:117] "RemoveContainer" containerID="7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" Oct 09 08:36:21 crc kubenswrapper[4872]: E1009 08:36:21.068159 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583\": container with ID starting with 7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583 not found: ID does not exist" containerID="7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.068188 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583"} err="failed to get container status \"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583\": rpc error: code = NotFound desc = could not find container \"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583\": container with ID starting with 7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583 not found: ID does not exist" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.068208 4872 scope.go:117] "RemoveContainer" containerID="034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.069208 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a"} err="failed to get container status \"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a\": rpc error: code = NotFound desc = could not find container \"034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a\": container with ID starting with 034c509558f5231fc76840c759097fb8244fdaf9a8d8b359d795dd51be36400a not found: ID does not exist" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.069233 4872 scope.go:117] "RemoveContainer" containerID="7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.083160 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5bf66bd496-2sdwc" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.083445 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583"} err="failed to get container status \"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583\": rpc error: code = NotFound desc = could not find container \"7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583\": container with ID starting with 7febb77a897e67671d5f669385733be6e35a1c09e7bcb1755e2f49bd1d22a583 not found: ID does not exist" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.087656 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.087697 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4866439-dbe8-4715-afda-5965d455af54-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.180184 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.302338 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.334014 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.381214 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:21 crc kubenswrapper[4872]: E1009 08:36:21.381788 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api-log" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.381817 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api-log" Oct 09 08:36:21 crc kubenswrapper[4872]: E1009 08:36:21.381868 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.381878 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.382070 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api-log" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.382088 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4866439-dbe8-4715-afda-5965d455af54" containerName="cinder-api" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.383782 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.398055 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.398338 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.398502 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.410561 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.435788 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514045 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqrsz\" (UniqueName: \"kubernetes.io/projected/10903b59-1315-45e1-b734-40c85cfa17e7-kube-api-access-qqrsz\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10903b59-1315-45e1-b734-40c85cfa17e7-logs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514166 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514203 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-scripts\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514248 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10903b59-1315-45e1-b734-40c85cfa17e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514302 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514331 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.514364 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616365 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616651 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10903b59-1315-45e1-b734-40c85cfa17e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616678 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616699 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616739 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616795 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/10903b59-1315-45e1-b734-40c85cfa17e7-etc-machine-id\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616841 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqrsz\" (UniqueName: \"kubernetes.io/projected/10903b59-1315-45e1-b734-40c85cfa17e7-kube-api-access-qqrsz\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616864 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10903b59-1315-45e1-b734-40c85cfa17e7-logs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616898 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.616922 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-scripts\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.618569 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10903b59-1315-45e1-b734-40c85cfa17e7-logs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.629623 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-public-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.629802 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data-custom\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.630469 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-scripts\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.632466 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-config-data\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.632912 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.633416 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10903b59-1315-45e1-b734-40c85cfa17e7-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.640251 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqrsz\" (UniqueName: \"kubernetes.io/projected/10903b59-1315-45e1-b734-40c85cfa17e7-kube-api-access-qqrsz\") pod \"cinder-api-0\" (UID: \"10903b59-1315-45e1-b734-40c85cfa17e7\") " pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.729529 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.949038 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerStarted","Data":"c8fd4d663efe03d4469617c4e555bd2c21a368771a032cc7e6dee83bea2f17c9"} Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.950680 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon-log" containerID="cri-o://648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70" gracePeriod=30 Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.951979 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.952029 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:21 crc kubenswrapper[4872]: I1009 08:36:21.952086 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" containerID="cri-o://61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7" gracePeriod=30 Oct 09 08:36:22 crc kubenswrapper[4872]: I1009 08:36:22.215466 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" podStartSLOduration=4.215441655 podStartE2EDuration="4.215441655s" podCreationTimestamp="2025-10-09 08:36:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:21.988759324 +0000 UTC m=+1140.179287960" watchObservedRunningTime="2025-10-09 08:36:22.215441655 +0000 UTC m=+1140.405970291" Oct 09 08:36:22 crc kubenswrapper[4872]: I1009 08:36:22.217160 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 09 08:36:22 crc kubenswrapper[4872]: W1009 08:36:22.231315 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10903b59_1315_45e1_b734_40c85cfa17e7.slice/crio-08cfb37cce89e6e33a8e2575cb2b693ac95841d38333cc954f73513d32d150e5 WatchSource:0}: Error finding container 08cfb37cce89e6e33a8e2575cb2b693ac95841d38333cc954f73513d32d150e5: Status 404 returned error can't find the container with id 08cfb37cce89e6e33a8e2575cb2b693ac95841d38333cc954f73513d32d150e5 Oct 09 08:36:22 crc kubenswrapper[4872]: I1009 08:36:22.498275 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4866439-dbe8-4715-afda-5965d455af54" path="/var/lib/kubelet/pods/e4866439-dbe8-4715-afda-5965d455af54/volumes" Oct 09 08:36:22 crc kubenswrapper[4872]: I1009 08:36:22.558974 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:36:22 crc kubenswrapper[4872]: I1009 08:36:22.988928 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerStarted","Data":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} Oct 09 08:36:23 crc kubenswrapper[4872]: I1009 08:36:23.008200 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10903b59-1315-45e1-b734-40c85cfa17e7","Type":"ContainerStarted","Data":"08cfb37cce89e6e33a8e2575cb2b693ac95841d38333cc954f73513d32d150e5"} Oct 09 08:36:23 crc kubenswrapper[4872]: I1009 08:36:23.694776 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:23 crc kubenswrapper[4872]: I1009 08:36:23.803037 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.020497 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10903b59-1315-45e1-b734-40c85cfa17e7","Type":"ContainerStarted","Data":"fcf223620889b7d0b0dab31ac112ce77f7f586f1ede55dc120123a994efc3c69"} Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.023566 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerStarted","Data":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.023600 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerStarted","Data":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.061011 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.133791 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.219574 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.219845 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="dnsmasq-dns" containerID="cri-o://d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23" gracePeriod=10 Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.407408 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.762139 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-c68d5ff89-tcgwk" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.774578 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.830464 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.830750 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66fb896578-k7hvx" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-api" containerID="cri-o://95252d2cac742c3f41f8739ec31c6f49eee975bf4092730720e7f3e348b30f02" gracePeriod=30 Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.831225 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-66fb896578-k7hvx" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-httpd" containerID="cri-o://d258faa64da10e60d56fe7f96c08d75962ba9be75b1fff2a7b65f225fad9a31b" gracePeriod=30 Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.894760 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7489b48876-4zz4g" Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.896690 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.896758 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.896827 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4kq5\" (UniqueName: \"kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.896915 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.897010 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.897093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config\") pod \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\" (UID: \"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c\") " Oct 09 08:36:24 crc kubenswrapper[4872]: I1009 08:36:24.921889 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5" (OuterVolumeSpecName: "kube-api-access-s4kq5") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "kube-api-access-s4kq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.005077 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4kq5\" (UniqueName: \"kubernetes.io/projected/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-kube-api-access-s4kq5\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.018208 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.048180 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config" (OuterVolumeSpecName: "config") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.051449 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.063197 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.100912 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"10903b59-1315-45e1-b734-40c85cfa17e7","Type":"ContainerStarted","Data":"89d41e737bee7880d112fca3115365540c0fdb1f249968f8b8b797e3ad9d4c45"} Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.102178 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.113850 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.113885 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.113895 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.113904 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.148854 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.14883157 podStartE2EDuration="4.14883157s" podCreationTimestamp="2025-10-09 08:36:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:25.1457189 +0000 UTC m=+1143.336247526" watchObservedRunningTime="2025-10-09 08:36:25.14883157 +0000 UTC m=+1143.339360206" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.156545 4872 generic.go:334] "Generic (PLEG): container finished" podID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerID="d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23" exitCode=0 Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.156773 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.157106 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" event={"ID":"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c","Type":"ContainerDied","Data":"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23"} Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.157188 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc5c4795-27zw7" event={"ID":"93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c","Type":"ContainerDied","Data":"b3cafa624f9ae998609735a16478372c05d9ade6977d3fee62dbfbd22c651a58"} Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.157207 4872 scope.go:117] "RemoveContainer" containerID="d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.173488 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" (UID: "93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.205167 4872 scope.go:117] "RemoveContainer" containerID="e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.217260 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.256983 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.349888 4872 scope.go:117] "RemoveContainer" containerID="d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23" Oct 09 08:36:25 crc kubenswrapper[4872]: E1009 08:36:25.350724 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23\": container with ID starting with d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23 not found: ID does not exist" containerID="d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.350763 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23"} err="failed to get container status \"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23\": rpc error: code = NotFound desc = could not find container \"d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23\": container with ID starting with d22e926c9e5e5c82f72cffc38f1809c87f28e5ed8fc5926bf867446ca91dbe23 not found: ID does not exist" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.350792 4872 scope.go:117] "RemoveContainer" containerID="e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee" Oct 09 08:36:25 crc kubenswrapper[4872]: E1009 08:36:25.351048 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee\": container with ID starting with e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee not found: ID does not exist" containerID="e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.351074 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee"} err="failed to get container status \"e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee\": rpc error: code = NotFound desc = could not find container \"e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee\": container with ID starting with e8ff19f5e7dfa9f569d5736f7d1467bab97aebdfc98f791c6e449841970081ee not found: ID does not exist" Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.499425 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.506114 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc5c4795-27zw7"] Oct 09 08:36:25 crc kubenswrapper[4872]: I1009 08:36:25.683559 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.051601 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 09 08:36:26 crc kubenswrapper[4872]: E1009 08:36:26.052081 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="dnsmasq-dns" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.052097 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="dnsmasq-dns" Oct 09 08:36:26 crc kubenswrapper[4872]: E1009 08:36:26.052125 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="init" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.052133 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="init" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.052362 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" containerName="dnsmasq-dns" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.053156 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.055583 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.055808 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qjmdm" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.056004 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.065089 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.159325 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.162678 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7vd8\" (UniqueName: \"kubernetes.io/projected/49ba1d52-eb34-49dd-8e46-34219f2e8336-kube-api-access-n7vd8\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.162743 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.162775 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.162859 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config-secret\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.187140 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerStarted","Data":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.188049 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.192165 4872 generic.go:334] "Generic (PLEG): container finished" podID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerID="61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7" exitCode=0 Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.192216 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerDied","Data":"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7"} Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.196076 4872 generic.go:334] "Generic (PLEG): container finished" podID="3f92a363-0e51-4968-bbf2-880a132a25db" containerID="d258faa64da10e60d56fe7f96c08d75962ba9be75b1fff2a7b65f225fad9a31b" exitCode=0 Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.196884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerDied","Data":"d258faa64da10e60d56fe7f96c08d75962ba9be75b1fff2a7b65f225fad9a31b"} Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.197036 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="cinder-scheduler" containerID="cri-o://aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e" gracePeriod=30 Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.197118 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="probe" containerID="cri-o://5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78" gracePeriod=30 Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.215291 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.438209614 podStartE2EDuration="6.215274036s" podCreationTimestamp="2025-10-09 08:36:20 +0000 UTC" firstStartedPulling="2025-10-09 08:36:21.415385167 +0000 UTC m=+1139.605913793" lastFinishedPulling="2025-10-09 08:36:25.192449589 +0000 UTC m=+1143.382978215" observedRunningTime="2025-10-09 08:36:26.211211228 +0000 UTC m=+1144.401739864" watchObservedRunningTime="2025-10-09 08:36:26.215274036 +0000 UTC m=+1144.405802692" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.264909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7vd8\" (UniqueName: \"kubernetes.io/projected/49ba1d52-eb34-49dd-8e46-34219f2e8336-kube-api-access-n7vd8\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.264984 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.265007 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.265111 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config-secret\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.266244 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.270364 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-openstack-config-secret\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.271067 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/49ba1d52-eb34-49dd-8e46-34219f2e8336-combined-ca-bundle\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.314259 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7vd8\" (UniqueName: \"kubernetes.io/projected/49ba1d52-eb34-49dd-8e46-34219f2e8336-kube-api-access-n7vd8\") pod \"openstackclient\" (UID: \"49ba1d52-eb34-49dd-8e46-34219f2e8336\") " pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.425083 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.477469 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c" path="/var/lib/kubelet/pods/93a8e711-2e0c-4cf5-b3ec-3fffb6f8581c/volumes" Oct 09 08:36:26 crc kubenswrapper[4872]: I1009 08:36:26.767256 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 09 08:36:26 crc kubenswrapper[4872]: W1009 08:36:26.780263 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod49ba1d52_eb34_49dd_8e46_34219f2e8336.slice/crio-b0a756d969ef8c8726b4561952e45de4c5264cba77e28e275dc6dd79001b9a82 WatchSource:0}: Error finding container b0a756d969ef8c8726b4561952e45de4c5264cba77e28e275dc6dd79001b9a82: Status 404 returned error can't find the container with id b0a756d969ef8c8726b4561952e45de4c5264cba77e28e275dc6dd79001b9a82 Oct 09 08:36:27 crc kubenswrapper[4872]: I1009 08:36:27.205295 4872 generic.go:334] "Generic (PLEG): container finished" podID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerID="5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78" exitCode=0 Oct 09 08:36:27 crc kubenswrapper[4872]: I1009 08:36:27.205372 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerDied","Data":"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78"} Oct 09 08:36:27 crc kubenswrapper[4872]: I1009 08:36:27.207008 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"49ba1d52-eb34-49dd-8e46-34219f2e8336","Type":"ContainerStarted","Data":"b0a756d969ef8c8726b4561952e45de4c5264cba77e28e275dc6dd79001b9a82"} Oct 09 08:36:28 crc kubenswrapper[4872]: I1009 08:36:28.635979 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58c4bcb5d6-tl5n8" Oct 09 08:36:28 crc kubenswrapper[4872]: I1009 08:36:28.713850 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:28 crc kubenswrapper[4872]: I1009 08:36:28.714076 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-846c9bf8fd-2l4kt" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api-log" containerID="cri-o://471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c" gracePeriod=30 Oct 09 08:36:28 crc kubenswrapper[4872]: I1009 08:36:28.721473 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-846c9bf8fd-2l4kt" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api" containerID="cri-o://197aaf63fc4c4afddb768ee10ed2e17d1d4983b192ac4fc05912e4cd5b72ef9b" gracePeriod=30 Oct 09 08:36:28 crc kubenswrapper[4872]: E1009 08:36:28.930590 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02d1b266_8d27_406e_8083_aa667f65b8ad.slice/crio-471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c.scope\": RecentStats: unable to find data in memory cache]" Oct 09 08:36:29 crc kubenswrapper[4872]: I1009 08:36:29.234999 4872 generic.go:334] "Generic (PLEG): container finished" podID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerID="471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c" exitCode=143 Oct 09 08:36:29 crc kubenswrapper[4872]: I1009 08:36:29.235044 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerDied","Data":"471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c"} Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.082373 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.087339 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.087370 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.088067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.088673 4872 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.189433 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.189616 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.189686 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wg2wh\" (UniqueName: \"kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.189713 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom\") pod \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\" (UID: \"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca\") " Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.194745 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.195067 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts" (OuterVolumeSpecName: "scripts") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.198130 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh" (OuterVolumeSpecName: "kube-api-access-wg2wh") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "kube-api-access-wg2wh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.260993 4872 generic.go:334] "Generic (PLEG): container finished" podID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerID="aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e" exitCode=0 Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.261033 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerDied","Data":"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e"} Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.261056 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1fcf1521-b4c8-4a2a-be8f-31c48c2960ca","Type":"ContainerDied","Data":"4453d37f68179950f835eb0a781feacfa69dd801f7e98e9fcf5e7943b3cb8b9f"} Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.261072 4872 scope.go:117] "RemoveContainer" containerID="5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.261175 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.261282 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data" (OuterVolumeSpecName: "config-data") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.275532 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" (UID: "1fcf1521-b4c8-4a2a-be8f-31c48c2960ca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.291186 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wg2wh\" (UniqueName: \"kubernetes.io/projected/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-kube-api-access-wg2wh\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.291217 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.291227 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.291235 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.291243 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.298124 4872 scope.go:117] "RemoveContainer" containerID="aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.324067 4872 scope.go:117] "RemoveContainer" containerID="5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78" Oct 09 08:36:31 crc kubenswrapper[4872]: E1009 08:36:31.324525 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78\": container with ID starting with 5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78 not found: ID does not exist" containerID="5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.324557 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78"} err="failed to get container status \"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78\": rpc error: code = NotFound desc = could not find container \"5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78\": container with ID starting with 5292717552c54878310a3ccdade419c71f2b5f26a455832a7088931460624e78 not found: ID does not exist" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.324578 4872 scope.go:117] "RemoveContainer" containerID="aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e" Oct 09 08:36:31 crc kubenswrapper[4872]: E1009 08:36:31.324876 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e\": container with ID starting with aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e not found: ID does not exist" containerID="aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.324899 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e"} err="failed to get container status \"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e\": rpc error: code = NotFound desc = could not find container \"aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e\": container with ID starting with aa8e182aced30ad9d85ed9d1e71343eb64b3c89f46ac8516b7e17e71b7ff349e not found: ID does not exist" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.628768 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.681698 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.699718 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:31 crc kubenswrapper[4872]: E1009 08:36:31.700177 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="probe" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.700201 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="probe" Oct 09 08:36:31 crc kubenswrapper[4872]: E1009 08:36:31.700228 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="cinder-scheduler" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.700234 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="cinder-scheduler" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.700406 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="probe" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.700423 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" containerName="cinder-scheduler" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.702265 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.704784 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.722684 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906624 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l884f\" (UniqueName: \"kubernetes.io/projected/b221b0fe-0514-409e-9ee6-966c6ceeb683-kube-api-access-l884f\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906706 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b221b0fe-0514-409e-9ee6-966c6ceeb683-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906771 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906795 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906833 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-scripts\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:31 crc kubenswrapper[4872]: I1009 08:36:31.906866 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.009772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-scripts\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.009865 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.009938 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l884f\" (UniqueName: \"kubernetes.io/projected/b221b0fe-0514-409e-9ee6-966c6ceeb683-kube-api-access-l884f\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.009970 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b221b0fe-0514-409e-9ee6-966c6ceeb683-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.010035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.010073 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.013935 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b221b0fe-0514-409e-9ee6-966c6ceeb683-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.020802 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-scripts\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.020994 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.021770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.028663 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b221b0fe-0514-409e-9ee6-966c6ceeb683-config-data\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.054414 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l884f\" (UniqueName: \"kubernetes.io/projected/b221b0fe-0514-409e-9ee6-966c6ceeb683-kube-api-access-l884f\") pod \"cinder-scheduler-0\" (UID: \"b221b0fe-0514-409e-9ee6-966c6ceeb683\") " pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.100272 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-846c9bf8fd-2l4kt" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.100520 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-846c9bf8fd-2l4kt" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.161:9311/healthcheck\": dial tcp 10.217.0.161:9311: connect: connection refused" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.328415 4872 generic.go:334] "Generic (PLEG): container finished" podID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerID="197aaf63fc4c4afddb768ee10ed2e17d1d4983b192ac4fc05912e4cd5b72ef9b" exitCode=0 Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.328457 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerDied","Data":"197aaf63fc4c4afddb768ee10ed2e17d1d4983b192ac4fc05912e4cd5b72ef9b"} Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.343086 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.390408 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-5b44c8599-9mlwc"] Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.392119 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.398971 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.399181 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.399280 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.399363 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b44c8599-9mlwc"] Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.446197 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.478959 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1fcf1521-b4c8-4a2a-be8f-31c48c2960ca" path="/var/lib/kubelet/pods/1fcf1521-b4c8-4a2a-be8f-31c48c2960ca/volumes" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.529915 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-etc-swift\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.529966 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-config-data\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.529995 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-internal-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.530038 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-log-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.530067 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmlvh\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-kube-api-access-hmlvh\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.530206 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-combined-ca-bundle\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.530242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-public-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.530303 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-run-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.546761 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.632260 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data\") pod \"02d1b266-8d27-406e-8083-aa667f65b8ad\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.632407 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle\") pod \"02d1b266-8d27-406e-8083-aa667f65b8ad\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.632434 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom\") pod \"02d1b266-8d27-406e-8083-aa667f65b8ad\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.632554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs\") pod \"02d1b266-8d27-406e-8083-aa667f65b8ad\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.632627 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5j2qh\" (UniqueName: \"kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh\") pod \"02d1b266-8d27-406e-8083-aa667f65b8ad\" (UID: \"02d1b266-8d27-406e-8083-aa667f65b8ad\") " Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.633180 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-log-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.633228 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmlvh\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-kube-api-access-hmlvh\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.633387 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-combined-ca-bundle\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.633689 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-public-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.634687 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-run-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.634794 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-etc-swift\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.634860 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-config-data\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.634894 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-internal-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.637228 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs" (OuterVolumeSpecName: "logs") pod "02d1b266-8d27-406e-8083-aa667f65b8ad" (UID: "02d1b266-8d27-406e-8083-aa667f65b8ad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.638193 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-run-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.638525 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dbc66806-aef2-4aa8-86ca-c48f90386f6c-log-httpd\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.653111 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh" (OuterVolumeSpecName: "kube-api-access-5j2qh") pod "02d1b266-8d27-406e-8083-aa667f65b8ad" (UID: "02d1b266-8d27-406e-8083-aa667f65b8ad"). InnerVolumeSpecName "kube-api-access-5j2qh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.653135 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "02d1b266-8d27-406e-8083-aa667f65b8ad" (UID: "02d1b266-8d27-406e-8083-aa667f65b8ad"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.661976 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-config-data\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.668798 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-etc-swift\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.670034 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-internal-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.675239 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-public-tls-certs\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.677090 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmlvh\" (UniqueName: \"kubernetes.io/projected/dbc66806-aef2-4aa8-86ca-c48f90386f6c-kube-api-access-hmlvh\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.698426 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc66806-aef2-4aa8-86ca-c48f90386f6c-combined-ca-bundle\") pod \"swift-proxy-5b44c8599-9mlwc\" (UID: \"dbc66806-aef2-4aa8-86ca-c48f90386f6c\") " pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.730807 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "02d1b266-8d27-406e-8083-aa667f65b8ad" (UID: "02d1b266-8d27-406e-8083-aa667f65b8ad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.742408 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.742436 4872 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.742445 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/02d1b266-8d27-406e-8083-aa667f65b8ad-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.742455 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5j2qh\" (UniqueName: \"kubernetes.io/projected/02d1b266-8d27-406e-8083-aa667f65b8ad-kube-api-access-5j2qh\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.776082 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.833609 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data" (OuterVolumeSpecName: "config-data") pod "02d1b266-8d27-406e-8083-aa667f65b8ad" (UID: "02d1b266-8d27-406e-8083-aa667f65b8ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:32 crc kubenswrapper[4872]: I1009 08:36:32.843547 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/02d1b266-8d27-406e-8083-aa667f65b8ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.017162 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-66db9d8d8d-cz7s2" Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.116113 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.378982 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b221b0fe-0514-409e-9ee6-966c6ceeb683","Type":"ContainerStarted","Data":"a9fff05ba08bef3c4792394dbe198a97bcb89e05820db1cacececf6ae80f8b24"} Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.383873 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-846c9bf8fd-2l4kt" event={"ID":"02d1b266-8d27-406e-8083-aa667f65b8ad","Type":"ContainerDied","Data":"a32606e1cf9b49455fdacb1ad04f68c74d0ca1c77436f1e12165f4bcf0f46b8b"} Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.383969 4872 scope.go:117] "RemoveContainer" containerID="197aaf63fc4c4afddb768ee10ed2e17d1d4983b192ac4fc05912e4cd5b72ef9b" Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.383905 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-846c9bf8fd-2l4kt" Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.384068 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.384451 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-central-agent" containerID="cri-o://bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" gracePeriod=30 Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.384503 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="proxy-httpd" containerID="cri-o://2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" gracePeriod=30 Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.384514 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-notification-agent" containerID="cri-o://36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" gracePeriod=30 Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.384488 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="sg-core" containerID="cri-o://f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" gracePeriod=30 Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.452824 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.462629 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-846c9bf8fd-2l4kt"] Oct 09 08:36:33 crc kubenswrapper[4872]: I1009 08:36:33.480252 4872 scope.go:117] "RemoveContainer" containerID="471a99dacb795d35c23969cc5314400552a3d94d0558f606f43ca2c84bf3985c" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.032805 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-5b44c8599-9mlwc"] Oct 09 08:36:34 crc kubenswrapper[4872]: W1009 08:36:34.040058 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbc66806_aef2_4aa8_86ca_c48f90386f6c.slice/crio-f553af00bd763c38e378cf77e4177dced746a50f898114b56dd27a04bbbc5bc2 WatchSource:0}: Error finding container f553af00bd763c38e378cf77e4177dced746a50f898114b56dd27a04bbbc5bc2: Status 404 returned error can't find the container with id f553af00bd763c38e378cf77e4177dced746a50f898114b56dd27a04bbbc5bc2 Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.378473 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.426059 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b221b0fe-0514-409e-9ee6-966c6ceeb683","Type":"ContainerStarted","Data":"7cb718c62ca93313a61ce0057df862d79686fea771c8ee12e5d78da61ae37860"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.427333 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b44c8599-9mlwc" event={"ID":"dbc66806-aef2-4aa8-86ca-c48f90386f6c","Type":"ContainerStarted","Data":"f553af00bd763c38e378cf77e4177dced746a50f898114b56dd27a04bbbc5bc2"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449458 4872 generic.go:334] "Generic (PLEG): container finished" podID="15813d5b-97ae-47f4-b672-3a715e096eef" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" exitCode=0 Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449517 4872 generic.go:334] "Generic (PLEG): container finished" podID="15813d5b-97ae-47f4-b672-3a715e096eef" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" exitCode=2 Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449509 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerDied","Data":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449572 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerDied","Data":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449529 4872 generic.go:334] "Generic (PLEG): container finished" podID="15813d5b-97ae-47f4-b672-3a715e096eef" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" exitCode=0 Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449598 4872 scope.go:117] "RemoveContainer" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449620 4872 generic.go:334] "Generic (PLEG): container finished" podID="15813d5b-97ae-47f4-b672-3a715e096eef" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" exitCode=0 Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449586 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerDied","Data":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449560 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449769 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerDied","Data":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.449797 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"15813d5b-97ae-47f4-b672-3a715e096eef","Type":"ContainerDied","Data":"c8fd4d663efe03d4469617c4e555bd2c21a368771a032cc7e6dee83bea2f17c9"} Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.481886 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" path="/var/lib/kubelet/pods/02d1b266-8d27-406e-8083-aa667f65b8ad/volumes" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.491582 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fz66z\" (UniqueName: \"kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.491912 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.491972 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.492039 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.492074 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.492107 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.492131 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd\") pod \"15813d5b-97ae-47f4-b672-3a715e096eef\" (UID: \"15813d5b-97ae-47f4-b672-3a715e096eef\") " Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.493727 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.494605 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.502802 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z" (OuterVolumeSpecName: "kube-api-access-fz66z") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "kube-api-access-fz66z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.502909 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts" (OuterVolumeSpecName: "scripts") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.545238 4872 scope.go:117] "RemoveContainer" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.594757 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.594790 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fz66z\" (UniqueName: \"kubernetes.io/projected/15813d5b-97ae-47f4-b672-3a715e096eef-kube-api-access-fz66z\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.594801 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.594810 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/15813d5b-97ae-47f4-b672-3a715e096eef-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.608622 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.611976 4872 scope.go:117] "RemoveContainer" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.673935 4872 scope.go:117] "RemoveContainer" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.700536 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.711369 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.805328 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.900288 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data" (OuterVolumeSpecName: "config-data") pod "15813d5b-97ae-47f4-b672-3a715e096eef" (UID: "15813d5b-97ae-47f4-b672-3a715e096eef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:34 crc kubenswrapper[4872]: I1009 08:36:34.907739 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15813d5b-97ae-47f4-b672-3a715e096eef-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.019159 4872 scope.go:117] "RemoveContainer" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.028027 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": container with ID starting with 2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf not found: ID does not exist" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.028105 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} err="failed to get container status \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": rpc error: code = NotFound desc = could not find container \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": container with ID starting with 2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.028160 4872 scope.go:117] "RemoveContainer" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.028763 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": container with ID starting with f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503 not found: ID does not exist" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.028792 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} err="failed to get container status \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": rpc error: code = NotFound desc = could not find container \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": container with ID starting with f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.028831 4872 scope.go:117] "RemoveContainer" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.030443 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": container with ID starting with 36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462 not found: ID does not exist" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.030501 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} err="failed to get container status \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": rpc error: code = NotFound desc = could not find container \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": container with ID starting with 36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.030534 4872 scope.go:117] "RemoveContainer" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.030944 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": container with ID starting with bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5 not found: ID does not exist" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.031004 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} err="failed to get container status \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": rpc error: code = NotFound desc = could not find container \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": container with ID starting with bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.031026 4872 scope.go:117] "RemoveContainer" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.031586 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} err="failed to get container status \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": rpc error: code = NotFound desc = could not find container \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": container with ID starting with 2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.031611 4872 scope.go:117] "RemoveContainer" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.031995 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} err="failed to get container status \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": rpc error: code = NotFound desc = could not find container \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": container with ID starting with f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.032048 4872 scope.go:117] "RemoveContainer" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.032482 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} err="failed to get container status \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": rpc error: code = NotFound desc = could not find container \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": container with ID starting with 36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.032511 4872 scope.go:117] "RemoveContainer" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.033143 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} err="failed to get container status \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": rpc error: code = NotFound desc = could not find container \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": container with ID starting with bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.033172 4872 scope.go:117] "RemoveContainer" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.035158 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} err="failed to get container status \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": rpc error: code = NotFound desc = could not find container \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": container with ID starting with 2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.035194 4872 scope.go:117] "RemoveContainer" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.035877 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} err="failed to get container status \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": rpc error: code = NotFound desc = could not find container \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": container with ID starting with f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.035926 4872 scope.go:117] "RemoveContainer" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.036457 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} err="failed to get container status \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": rpc error: code = NotFound desc = could not find container \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": container with ID starting with 36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.036504 4872 scope.go:117] "RemoveContainer" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.037547 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} err="failed to get container status \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": rpc error: code = NotFound desc = could not find container \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": container with ID starting with bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.037595 4872 scope.go:117] "RemoveContainer" containerID="2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.037910 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf"} err="failed to get container status \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": rpc error: code = NotFound desc = could not find container \"2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf\": container with ID starting with 2cf2f3a0d57fe6c0de67e658533047f1b27f4d88abc80011d2b2b856603972bf not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.037935 4872 scope.go:117] "RemoveContainer" containerID="f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.038543 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503"} err="failed to get container status \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": rpc error: code = NotFound desc = could not find container \"f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503\": container with ID starting with f69da557d1c4f6887655a066d435d321001aa485206d181882265074aee32503 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.038587 4872 scope.go:117] "RemoveContainer" containerID="36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.039628 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462"} err="failed to get container status \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": rpc error: code = NotFound desc = could not find container \"36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462\": container with ID starting with 36798ef672cd268c9933c57d71cf20b84b2e6e960c3594a75af7b6a56fd5a462 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.039681 4872 scope.go:117] "RemoveContainer" containerID="bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.040092 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5"} err="failed to get container status \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": rpc error: code = NotFound desc = could not find container \"bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5\": container with ID starting with bfa39e19679df8cdf6d3914afc2059146676cc7747af879b91bda2a1d1869bf5 not found: ID does not exist" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.089516 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.104710 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122072 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122516 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122539 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122557 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-central-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122566 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-central-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122612 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-notification-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122621 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-notification-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122652 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api-log" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122661 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api-log" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122674 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="proxy-httpd" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122684 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="proxy-httpd" Oct 09 08:36:35 crc kubenswrapper[4872]: E1009 08:36:35.122696 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="sg-core" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122704 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="sg-core" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122905 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="sg-core" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122931 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-central-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122949 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="ceilometer-notification-agent" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122965 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api-log" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122981 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="02d1b266-8d27-406e-8083-aa667f65b8ad" containerName="barbican-api" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.122994 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" containerName="proxy-httpd" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.125098 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.130569 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.130957 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.142847 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.220953 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.220992 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.221053 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.221124 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.221145 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.221176 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.221216 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwq44\" (UniqueName: \"kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.226108 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323009 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323045 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323070 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323095 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwq44\" (UniqueName: \"kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323142 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.323238 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.324574 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.329517 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.330138 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.334665 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.337494 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.346752 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.349391 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwq44\" (UniqueName: \"kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44\") pod \"ceilometer-0\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.450341 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.511589 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b44c8599-9mlwc" event={"ID":"dbc66806-aef2-4aa8-86ca-c48f90386f6c","Type":"ContainerStarted","Data":"faac0e8c1ebe740ec0b71af1ce7293dcdaa815737365f63d1fd752f95a97cdcf"} Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.511916 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-5b44c8599-9mlwc" event={"ID":"dbc66806-aef2-4aa8-86ca-c48f90386f6c","Type":"ContainerStarted","Data":"9e6eda391b5efda4c18960786e0653e9fde210c7d7a8ad73a44fd819e5976fc4"} Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.513465 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.513746 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.545785 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"b221b0fe-0514-409e-9ee6-966c6ceeb683","Type":"ContainerStarted","Data":"7225716cc1bc35ed4154470b4a678d7e0cf5a35b929ca8dbf533f7e262d620e4"} Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.551470 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-5b44c8599-9mlwc" podStartSLOduration=3.551446458 podStartE2EDuration="3.551446458s" podCreationTimestamp="2025-10-09 08:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:35.537632129 +0000 UTC m=+1153.728160755" watchObservedRunningTime="2025-10-09 08:36:35.551446458 +0000 UTC m=+1153.741975074" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.589179 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.589150916 podStartE2EDuration="4.589150916s" podCreationTimestamp="2025-10-09 08:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:35.571581969 +0000 UTC m=+1153.762110585" watchObservedRunningTime="2025-10-09 08:36:35.589150916 +0000 UTC m=+1153.779679542" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.684619 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 09 08:36:35 crc kubenswrapper[4872]: I1009 08:36:35.955818 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:36 crc kubenswrapper[4872]: I1009 08:36:36.474390 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15813d5b-97ae-47f4-b672-3a715e096eef" path="/var/lib/kubelet/pods/15813d5b-97ae-47f4-b672-3a715e096eef/volumes" Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.281612 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.282237 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-log" containerID="cri-o://95bb1ad59c73749dba34c8eb0bf532fc6ef770f520d61395aa50daf85ca92446" gracePeriod=30 Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.282338 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-httpd" containerID="cri-o://fdf21793e2e99ec9ff1aabbb3d51e6b0ba343ff4ca8b086f7944229443fd5b88" gracePeriod=30 Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.343679 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.580550 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a688f76-73bf-4be3-9665-26151a8749b1" containerID="95bb1ad59c73749dba34c8eb0bf532fc6ef770f520d61395aa50daf85ca92446" exitCode=143 Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.580667 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerDied","Data":"95bb1ad59c73749dba34c8eb0bf532fc6ef770f520d61395aa50daf85ca92446"} Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.586610 4872 generic.go:334] "Generic (PLEG): container finished" podID="3f92a363-0e51-4968-bbf2-880a132a25db" containerID="95252d2cac742c3f41f8739ec31c6f49eee975bf4092730720e7f3e348b30f02" exitCode=0 Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.586687 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerDied","Data":"95252d2cac742c3f41f8739ec31c6f49eee975bf4092730720e7f3e348b30f02"} Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.960412 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.960663 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-log" containerID="cri-o://42aac7f2fd68b309be6ee6a3668a30de61cf6ee3f575a160f926b70613f1ce8c" gracePeriod=30 Oct 09 08:36:37 crc kubenswrapper[4872]: I1009 08:36:37.960795 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-httpd" containerID="cri-o://ecdd549b08379bea6014ca0133872d82b31cc5ca73b37ca8701c2b32fa88291d" gracePeriod=30 Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.156655 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.156701 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.156740 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.157441 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.157491 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451" gracePeriod=600 Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.479623 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-kl2lk"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.482524 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.491736 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kl2lk"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.591375 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-wcspf"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.594010 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.615795 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wcspf"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.616700 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr8c6\" (UniqueName: \"kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6\") pod \"nova-api-db-create-kl2lk\" (UID: \"aed630ee-8d81-4c05-8cd0-85245b99fa00\") " pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.618508 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451" exitCode=0 Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.618566 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451"} Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.618595 4872 scope.go:117] "RemoveContainer" containerID="75c156b56bd66aa8cf30cadd1002243f41417fee084aa8c7c31d7c9507ec5e41" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.621401 4872 generic.go:334] "Generic (PLEG): container finished" podID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerID="42aac7f2fd68b309be6ee6a3668a30de61cf6ee3f575a160f926b70613f1ce8c" exitCode=143 Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.621423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerDied","Data":"42aac7f2fd68b309be6ee6a3668a30de61cf6ee3f575a160f926b70613f1ce8c"} Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.691808 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-dbs6x"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.693189 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.707855 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dbs6x"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.718164 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g9vc\" (UniqueName: \"kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc\") pod \"nova-cell0-db-create-wcspf\" (UID: \"896adf7e-0f68-4f60-bd0c-6e507673c221\") " pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.718377 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr8c6\" (UniqueName: \"kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6\") pod \"nova-api-db-create-kl2lk\" (UID: \"aed630ee-8d81-4c05-8cd0-85245b99fa00\") " pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.738765 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr8c6\" (UniqueName: \"kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6\") pod \"nova-api-db-create-kl2lk\" (UID: \"aed630ee-8d81-4c05-8cd0-85245b99fa00\") " pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.808316 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.821005 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvrcs\" (UniqueName: \"kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs\") pod \"nova-cell1-db-create-dbs6x\" (UID: \"a0703df6-2aa8-4de5-b54e-850b63effded\") " pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.821166 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g9vc\" (UniqueName: \"kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc\") pod \"nova-cell0-db-create-wcspf\" (UID: \"896adf7e-0f68-4f60-bd0c-6e507673c221\") " pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.853318 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.865172 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g9vc\" (UniqueName: \"kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc\") pod \"nova-cell0-db-create-wcspf\" (UID: \"896adf7e-0f68-4f60-bd0c-6e507673c221\") " pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.917084 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.923428 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvrcs\" (UniqueName: \"kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs\") pod \"nova-cell1-db-create-dbs6x\" (UID: \"a0703df6-2aa8-4de5-b54e-850b63effded\") " pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:38 crc kubenswrapper[4872]: I1009 08:36:38.938508 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvrcs\" (UniqueName: \"kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs\") pod \"nova-cell1-db-create-dbs6x\" (UID: \"a0703df6-2aa8-4de5-b54e-850b63effded\") " pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:39 crc kubenswrapper[4872]: I1009 08:36:39.017727 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:40 crc kubenswrapper[4872]: I1009 08:36:40.665231 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a688f76-73bf-4be3-9665-26151a8749b1" containerID="fdf21793e2e99ec9ff1aabbb3d51e6b0ba343ff4ca8b086f7944229443fd5b88" exitCode=0 Oct 09 08:36:40 crc kubenswrapper[4872]: I1009 08:36:40.665543 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerDied","Data":"fdf21793e2e99ec9ff1aabbb3d51e6b0ba343ff4ca8b086f7944229443fd5b88"} Oct 09 08:36:41 crc kubenswrapper[4872]: I1009 08:36:41.676264 4872 generic.go:334] "Generic (PLEG): container finished" podID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerID="ecdd549b08379bea6014ca0133872d82b31cc5ca73b37ca8701c2b32fa88291d" exitCode=0 Oct 09 08:36:41 crc kubenswrapper[4872]: I1009 08:36:41.676304 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerDied","Data":"ecdd549b08379bea6014ca0133872d82b31cc5ca73b37ca8701c2b32fa88291d"} Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.644703 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.661075 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.688796 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-66fb896578-k7hvx" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.689241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-66fb896578-k7hvx" event={"ID":"3f92a363-0e51-4968-bbf2-880a132a25db","Type":"ContainerDied","Data":"4169a26ade0e61ba21af3d85ff236c1ca06bc83d97c70860032741af62520130"} Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.689282 4872 scope.go:117] "RemoveContainer" containerID="d258faa64da10e60d56fe7f96c08d75962ba9be75b1fff2a7b65f225fad9a31b" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.697338 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f"} Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.700929 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerStarted","Data":"71b761c8d36d2654889ebbaaa26f3cab8b464b17209fbf16310e582cf94a0a01"} Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.716283 4872 scope.go:117] "RemoveContainer" containerID="95252d2cac742c3f41f8739ec31c6f49eee975bf4092730720e7f3e348b30f02" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.786089 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.787946 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-5b44c8599-9mlwc" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.814308 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mn7p\" (UniqueName: \"kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p\") pod \"3f92a363-0e51-4968-bbf2-880a132a25db\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.815548 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs\") pod \"3f92a363-0e51-4968-bbf2-880a132a25db\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.815585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config\") pod \"3f92a363-0e51-4968-bbf2-880a132a25db\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.815670 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle\") pod \"3f92a363-0e51-4968-bbf2-880a132a25db\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.815705 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config\") pod \"3f92a363-0e51-4968-bbf2-880a132a25db\" (UID: \"3f92a363-0e51-4968-bbf2-880a132a25db\") " Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.853054 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p" (OuterVolumeSpecName: "kube-api-access-9mn7p") pod "3f92a363-0e51-4968-bbf2-880a132a25db" (UID: "3f92a363-0e51-4968-bbf2-880a132a25db"). InnerVolumeSpecName "kube-api-access-9mn7p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.853367 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "3f92a363-0e51-4968-bbf2-880a132a25db" (UID: "3f92a363-0e51-4968-bbf2-880a132a25db"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.914624 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f92a363-0e51-4968-bbf2-880a132a25db" (UID: "3f92a363-0e51-4968-bbf2-880a132a25db"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.918378 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mn7p\" (UniqueName: \"kubernetes.io/projected/3f92a363-0e51-4968-bbf2-880a132a25db-kube-api-access-9mn7p\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.918401 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.918411 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.934472 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config" (OuterVolumeSpecName: "config") pod "3f92a363-0e51-4968-bbf2-880a132a25db" (UID: "3f92a363-0e51-4968-bbf2-880a132a25db"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.935159 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:36:42 crc kubenswrapper[4872]: I1009 08:36:42.968395 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "3f92a363-0e51-4968-bbf2-880a132a25db" (UID: "3f92a363-0e51-4968-bbf2-880a132a25db"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019168 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019225 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019280 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019363 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019397 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b82gc\" (UniqueName: \"kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019440 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019470 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019499 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs\") pod \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\" (UID: \"68d1d4be-05fb-4c0e-be1e-0f7293280b33\") " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.019730 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.020034 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.020054 4872 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.020067 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3f92a363-0e51-4968-bbf2-880a132a25db-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.027189 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs" (OuterVolumeSpecName: "logs") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.027713 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts" (OuterVolumeSpecName: "scripts") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.046898 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.057866 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc" (OuterVolumeSpecName: "kube-api-access-b82gc") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "kube-api-access-b82gc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.060840 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.067756 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.076445 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-66fb896578-k7hvx"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.095900 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-kl2lk"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.115754 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121311 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121335 4872 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121345 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b82gc\" (UniqueName: \"kubernetes.io/projected/68d1d4be-05fb-4c0e-be1e-0f7293280b33-kube-api-access-b82gc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121366 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121377 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.121386 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68d1d4be-05fb-4c0e-be1e-0f7293280b33-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.142706 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data" (OuterVolumeSpecName: "config-data") pod "68d1d4be-05fb-4c0e-be1e-0f7293280b33" (UID: "68d1d4be-05fb-4c0e-be1e-0f7293280b33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.169178 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Oct 09 08:36:43 crc kubenswrapper[4872]: W1009 08:36:43.185119 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda0703df6_2aa8_4de5_b54e_850b63effded.slice/crio-deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b WatchSource:0}: Error finding container deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b: Status 404 returned error can't find the container with id deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b Oct 09 08:36:43 crc kubenswrapper[4872]: W1009 08:36:43.198727 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod896adf7e_0f68_4f60_bd0c_6e507673c221.slice/crio-c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f WatchSource:0}: Error finding container c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f: Status 404 returned error can't find the container with id c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.222418 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-dbs6x"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.227492 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.227544 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68d1d4be-05fb-4c0e-be1e-0f7293280b33-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.266140 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-wcspf"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.720174 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"68d1d4be-05fb-4c0e-be1e-0f7293280b33","Type":"ContainerDied","Data":"4e8701d58d65de16e060f9fd2ed02e7222dc56842d04161220209481a7845e2f"} Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.721107 4872 scope.go:117] "RemoveContainer" containerID="ecdd549b08379bea6014ca0133872d82b31cc5ca73b37ca8701c2b32fa88291d" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.720280 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.723470 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wcspf" event={"ID":"896adf7e-0f68-4f60-bd0c-6e507673c221","Type":"ContainerStarted","Data":"c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f"} Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.729542 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kl2lk" event={"ID":"aed630ee-8d81-4c05-8cd0-85245b99fa00","Type":"ContainerStarted","Data":"f71403cf831e032254cc9c038137ea151e30a36e84e5e734c6a8e7258fabc94f"} Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.735760 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbs6x" event={"ID":"a0703df6-2aa8-4de5-b54e-850b63effded","Type":"ContainerStarted","Data":"deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b"} Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.740441 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"49ba1d52-eb34-49dd-8e46-34219f2e8336","Type":"ContainerStarted","Data":"0cd78cfba361bf3901848799cc7f3c1a56ab3f838aaef8db0d6c4318558d5689"} Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.759699 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.106801784 podStartE2EDuration="17.759678049s" podCreationTimestamp="2025-10-09 08:36:26 +0000 UTC" firstStartedPulling="2025-10-09 08:36:26.782416073 +0000 UTC m=+1144.972944699" lastFinishedPulling="2025-10-09 08:36:42.435292338 +0000 UTC m=+1160.625820964" observedRunningTime="2025-10-09 08:36:43.755831908 +0000 UTC m=+1161.946360534" watchObservedRunningTime="2025-10-09 08:36:43.759678049 +0000 UTC m=+1161.950206675" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.784376 4872 scope.go:117] "RemoveContainer" containerID="42aac7f2fd68b309be6ee6a3668a30de61cf6ee3f575a160f926b70613f1ce8c" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.800058 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.832014 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.856472 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:43 crc kubenswrapper[4872]: E1009 08:36:43.857005 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857026 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: E1009 08:36:43.857051 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-api" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857059 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-api" Oct 09 08:36:43 crc kubenswrapper[4872]: E1009 08:36:43.857087 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-log" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857095 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-log" Oct 09 08:36:43 crc kubenswrapper[4872]: E1009 08:36:43.857114 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857122 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857332 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857357 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-httpd" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857371 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" containerName="neutron-api" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.857392 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" containerName="glance-log" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.865224 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.869997 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.870734 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.878604 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.914674 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.943871 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gplgw\" (UniqueName: \"kubernetes.io/projected/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-kube-api-access-gplgw\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.943938 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.943985 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.944014 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.944041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-config-data\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.944079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-scripts\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.944113 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-logs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:43 crc kubenswrapper[4872]: I1009 08:36:43.944133 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.045234 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.045978 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046003 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046064 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046147 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046213 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046237 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046358 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cgmw\" (UniqueName: \"kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw\") pod \"2a688f76-73bf-4be3-9665-26151a8749b1\" (UID: \"2a688f76-73bf-4be3-9665-26151a8749b1\") " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046616 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046674 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046704 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-config-data\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046761 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-scripts\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046808 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-logs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046830 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046862 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gplgw\" (UniqueName: \"kubernetes.io/projected/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-kube-api-access-gplgw\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.046907 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.047401 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.048053 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.048671 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.049851 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs" (OuterVolumeSpecName: "logs") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.050368 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-logs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.069498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.074988 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.082425 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.083473 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-scripts\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.085797 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gplgw\" (UniqueName: \"kubernetes.io/projected/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-kube-api-access-gplgw\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.091218 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts" (OuterVolumeSpecName: "scripts") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.093332 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d651cc86-02b8-4d90-b4a0-3d1a2de1e300-config-data\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.094103 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw" (OuterVolumeSpecName: "kube-api-access-8cgmw") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "kube-api-access-8cgmw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.106258 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"d651cc86-02b8-4d90-b4a0-3d1a2de1e300\") " pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.118537 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149107 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149139 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cgmw\" (UniqueName: \"kubernetes.io/projected/2a688f76-73bf-4be3-9665-26151a8749b1-kube-api-access-8cgmw\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149150 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149159 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149187 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.149197 4872 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2a688f76-73bf-4be3-9665-26151a8749b1-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.157490 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.160731 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data" (OuterVolumeSpecName: "config-data") pod "2a688f76-73bf-4be3-9665-26151a8749b1" (UID: "2a688f76-73bf-4be3-9665-26151a8749b1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.183294 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.201060 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.250848 4872 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.250885 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.250897 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a688f76-73bf-4be3-9665-26151a8749b1-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.473490 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f92a363-0e51-4968-bbf2-880a132a25db" path="/var/lib/kubelet/pods/3f92a363-0e51-4968-bbf2-880a132a25db/volumes" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.477577 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68d1d4be-05fb-4c0e-be1e-0f7293280b33" path="/var/lib/kubelet/pods/68d1d4be-05fb-4c0e-be1e-0f7293280b33/volumes" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.750587 4872 generic.go:334] "Generic (PLEG): container finished" podID="a0703df6-2aa8-4de5-b54e-850b63effded" containerID="b1066cfed7505454036e5cc7ea38daf5d26a5c1bdb956384641b90fbd2390dea" exitCode=0 Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.750691 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbs6x" event={"ID":"a0703df6-2aa8-4de5-b54e-850b63effded","Type":"ContainerDied","Data":"b1066cfed7505454036e5cc7ea38daf5d26a5c1bdb956384641b90fbd2390dea"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.752593 4872 generic.go:334] "Generic (PLEG): container finished" podID="aed630ee-8d81-4c05-8cd0-85245b99fa00" containerID="e7997ef6320771fffc338d4411410a8304acf25cf8979aff8fd22d179dff1e5e" exitCode=0 Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.752662 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kl2lk" event={"ID":"aed630ee-8d81-4c05-8cd0-85245b99fa00","Type":"ContainerDied","Data":"e7997ef6320771fffc338d4411410a8304acf25cf8979aff8fd22d179dff1e5e"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.755940 4872 generic.go:334] "Generic (PLEG): container finished" podID="896adf7e-0f68-4f60-bd0c-6e507673c221" containerID="6de5737ca4cd4443921fe5b3136243b0206ba4c3da09e6a607757b640233d327" exitCode=0 Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.756021 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wcspf" event={"ID":"896adf7e-0f68-4f60-bd0c-6e507673c221","Type":"ContainerDied","Data":"6de5737ca4cd4443921fe5b3136243b0206ba4c3da09e6a607757b640233d327"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.758332 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerStarted","Data":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.758374 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerStarted","Data":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.760815 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"2a688f76-73bf-4be3-9665-26151a8749b1","Type":"ContainerDied","Data":"57bc46d3d46ca86bd57cf6caf3485ce07f16cf88947295820b44cc3f1b2fa7be"} Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.760846 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.760888 4872 scope.go:117] "RemoveContainer" containerID="fdf21793e2e99ec9ff1aabbb3d51e6b0ba343ff4ca8b086f7944229443fd5b88" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.790517 4872 scope.go:117] "RemoveContainer" containerID="95bb1ad59c73749dba34c8eb0bf532fc6ef770f520d61395aa50daf85ca92446" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.831920 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.864909 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.875959 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.909883 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:44 crc kubenswrapper[4872]: E1009 08:36:44.910633 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-httpd" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.910666 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-httpd" Oct 09 08:36:44 crc kubenswrapper[4872]: E1009 08:36:44.910720 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-log" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.910727 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-log" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.911451 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-httpd" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.911504 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" containerName="glance-log" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.913493 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.916511 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.916857 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.917920 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.970525 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4hcq\" (UniqueName: \"kubernetes.io/projected/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-kube-api-access-j4hcq\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.970995 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971066 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971092 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971127 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-logs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971155 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:44 crc kubenswrapper[4872]: I1009 08:36:44.971275 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072362 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4hcq\" (UniqueName: \"kubernetes.io/projected/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-kube-api-access-j4hcq\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072416 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072445 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072462 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072482 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072507 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-logs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072527 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.072579 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.073059 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.074059 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-logs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.074038 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.081202 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.083344 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.087119 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.088734 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.091180 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4hcq\" (UniqueName: \"kubernetes.io/projected/a70f38a0-6bd6-45c9-8dc6-6f540842a4be-kube-api-access-j4hcq\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.105427 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-internal-api-0\" (UID: \"a70f38a0-6bd6-45c9-8dc6-6f540842a4be\") " pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.258336 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.683261 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-6cdd4b976-7cgml" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.145:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.145:8443: connect: connection refused" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.683810 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.797684 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerStarted","Data":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.809733 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d651cc86-02b8-4d90-b4a0-3d1a2de1e300","Type":"ContainerStarted","Data":"655814e5d65133c4fc3733f51f53ac64b01c35ade53a758708288675053faf5b"} Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.809776 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d651cc86-02b8-4d90-b4a0-3d1a2de1e300","Type":"ContainerStarted","Data":"168308ce226785b6a727396ddf71b273246dd23e38dc5346480068521fd373bb"} Oct 09 08:36:45 crc kubenswrapper[4872]: I1009 08:36:45.867418 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 09 08:36:45 crc kubenswrapper[4872]: W1009 08:36:45.880589 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda70f38a0_6bd6_45c9_8dc6_6f540842a4be.slice/crio-23f15822375c68a023e8f6c68818121b9332df5106c598d15b919e5e1abccdcb WatchSource:0}: Error finding container 23f15822375c68a023e8f6c68818121b9332df5106c598d15b919e5e1abccdcb: Status 404 returned error can't find the container with id 23f15822375c68a023e8f6c68818121b9332df5106c598d15b919e5e1abccdcb Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.370213 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.380955 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.406893 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.429505 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g9vc\" (UniqueName: \"kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc\") pod \"896adf7e-0f68-4f60-bd0c-6e507673c221\" (UID: \"896adf7e-0f68-4f60-bd0c-6e507673c221\") " Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.446882 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc" (OuterVolumeSpecName: "kube-api-access-9g9vc") pod "896adf7e-0f68-4f60-bd0c-6e507673c221" (UID: "896adf7e-0f68-4f60-bd0c-6e507673c221"). InnerVolumeSpecName "kube-api-access-9g9vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.482787 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a688f76-73bf-4be3-9665-26151a8749b1" path="/var/lib/kubelet/pods/2a688f76-73bf-4be3-9665-26151a8749b1/volumes" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.531334 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qr8c6\" (UniqueName: \"kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6\") pod \"aed630ee-8d81-4c05-8cd0-85245b99fa00\" (UID: \"aed630ee-8d81-4c05-8cd0-85245b99fa00\") " Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.531497 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvrcs\" (UniqueName: \"kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs\") pod \"a0703df6-2aa8-4de5-b54e-850b63effded\" (UID: \"a0703df6-2aa8-4de5-b54e-850b63effded\") " Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.532008 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g9vc\" (UniqueName: \"kubernetes.io/projected/896adf7e-0f68-4f60-bd0c-6e507673c221-kube-api-access-9g9vc\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.535626 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6" (OuterVolumeSpecName: "kube-api-access-qr8c6") pod "aed630ee-8d81-4c05-8cd0-85245b99fa00" (UID: "aed630ee-8d81-4c05-8cd0-85245b99fa00"). InnerVolumeSpecName "kube-api-access-qr8c6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.538840 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs" (OuterVolumeSpecName: "kube-api-access-hvrcs") pod "a0703df6-2aa8-4de5-b54e-850b63effded" (UID: "a0703df6-2aa8-4de5-b54e-850b63effded"). InnerVolumeSpecName "kube-api-access-hvrcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.634713 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qr8c6\" (UniqueName: \"kubernetes.io/projected/aed630ee-8d81-4c05-8cd0-85245b99fa00-kube-api-access-qr8c6\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.634742 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvrcs\" (UniqueName: \"kubernetes.io/projected/a0703df6-2aa8-4de5-b54e-850b63effded-kube-api-access-hvrcs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.830440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d651cc86-02b8-4d90-b4a0-3d1a2de1e300","Type":"ContainerStarted","Data":"074bbfa3454ea592506d5b0341aa0d96365697f6f719b06dcc4e897ab6d6cfea"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.833557 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-dbs6x" event={"ID":"a0703df6-2aa8-4de5-b54e-850b63effded","Type":"ContainerDied","Data":"deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.833596 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="deb3abec627c609103cce8aeda5cc9e2d9c78c23bcf2008494b5c8d48db5993b" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.833591 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-dbs6x" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.835651 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-kl2lk" event={"ID":"aed630ee-8d81-4c05-8cd0-85245b99fa00","Type":"ContainerDied","Data":"f71403cf831e032254cc9c038137ea151e30a36e84e5e734c6a8e7258fabc94f"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.835674 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-kl2lk" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.835711 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f71403cf831e032254cc9c038137ea151e30a36e84e5e734c6a8e7258fabc94f" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.838038 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-wcspf" event={"ID":"896adf7e-0f68-4f60-bd0c-6e507673c221","Type":"ContainerDied","Data":"c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.838076 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4228fe43964f441a219d83f078cb21a4d8c322396678a9a0fa084a4fa672d2f" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.838130 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-wcspf" Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.840825 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a70f38a0-6bd6-45c9-8dc6-6f540842a4be","Type":"ContainerStarted","Data":"2bb70fab50de10065b69a2c05afcdc088b3f38fa8aad5e8320b4946aa0ad9e95"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.840860 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a70f38a0-6bd6-45c9-8dc6-6f540842a4be","Type":"ContainerStarted","Data":"23f15822375c68a023e8f6c68818121b9332df5106c598d15b919e5e1abccdcb"} Oct 09 08:36:46 crc kubenswrapper[4872]: I1009 08:36:46.862952 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.862933705 podStartE2EDuration="3.862933705s" podCreationTimestamp="2025-10-09 08:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:46.85444375 +0000 UTC m=+1165.044972376" watchObservedRunningTime="2025-10-09 08:36:46.862933705 +0000 UTC m=+1165.053462331" Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.851680 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"a70f38a0-6bd6-45c9-8dc6-6f540842a4be","Type":"ContainerStarted","Data":"ae748a640d342425306b21a78ee6a004285df5d0cd9ec23e0cb0cf927e3b6f80"} Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.856956 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-central-agent" containerID="cri-o://0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" gracePeriod=30 Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.857151 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerStarted","Data":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.857193 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.857229 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="proxy-httpd" containerID="cri-o://1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" gracePeriod=30 Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.857269 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="sg-core" containerID="cri-o://536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" gracePeriod=30 Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.857304 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-notification-agent" containerID="cri-o://91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" gracePeriod=30 Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.877984 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.8779616900000002 podStartE2EDuration="3.87796169s" podCreationTimestamp="2025-10-09 08:36:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:36:47.874606544 +0000 UTC m=+1166.065135190" watchObservedRunningTime="2025-10-09 08:36:47.87796169 +0000 UTC m=+1166.068490316" Oct 09 08:36:47 crc kubenswrapper[4872]: I1009 08:36:47.904873 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.048390134 podStartE2EDuration="12.904842936s" podCreationTimestamp="2025-10-09 08:36:35 +0000 UTC" firstStartedPulling="2025-10-09 08:36:42.149869341 +0000 UTC m=+1160.340397967" lastFinishedPulling="2025-10-09 08:36:47.006322143 +0000 UTC m=+1165.196850769" observedRunningTime="2025-10-09 08:36:47.896346241 +0000 UTC m=+1166.086874897" watchObservedRunningTime="2025-10-09 08:36:47.904842936 +0000 UTC m=+1166.095371602" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.641197 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768453 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768503 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768537 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwq44\" (UniqueName: \"kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768663 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768728 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768787 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.768816 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd\") pod \"a75b743a-2d4c-4518-a829-71ae90478f45\" (UID: \"a75b743a-2d4c-4518-a829-71ae90478f45\") " Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.769724 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.771519 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.776245 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts" (OuterVolumeSpecName: "scripts") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.781841 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44" (OuterVolumeSpecName: "kube-api-access-hwq44") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "kube-api-access-hwq44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.813812 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.866886 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871025 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871061 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871076 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871089 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871101 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwq44\" (UniqueName: \"kubernetes.io/projected/a75b743a-2d4c-4518-a829-71ae90478f45-kube-api-access-hwq44\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871113 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/a75b743a-2d4c-4518-a829-71ae90478f45-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.871891 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data" (OuterVolumeSpecName: "config-data") pod "a75b743a-2d4c-4518-a829-71ae90478f45" (UID: "a75b743a-2d4c-4518-a829-71ae90478f45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.888743 4872 generic.go:334] "Generic (PLEG): container finished" podID="a75b743a-2d4c-4518-a829-71ae90478f45" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" exitCode=0 Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.888785 4872 generic.go:334] "Generic (PLEG): container finished" podID="a75b743a-2d4c-4518-a829-71ae90478f45" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" exitCode=2 Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.888794 4872 generic.go:334] "Generic (PLEG): container finished" podID="a75b743a-2d4c-4518-a829-71ae90478f45" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" exitCode=0 Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.888803 4872 generic.go:334] "Generic (PLEG): container finished" podID="a75b743a-2d4c-4518-a829-71ae90478f45" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" exitCode=0 Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889094 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889163 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerDied","Data":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889225 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerDied","Data":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889241 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerDied","Data":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889253 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerDied","Data":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889306 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"a75b743a-2d4c-4518-a829-71ae90478f45","Type":"ContainerDied","Data":"71b761c8d36d2654889ebbaaa26f3cab8b464b17209fbf16310e582cf94a0a01"} Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.889284 4872 scope.go:117] "RemoveContainer" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.911140 4872 scope.go:117] "RemoveContainer" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.930712 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.943558 4872 scope.go:117] "RemoveContainer" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.949347 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.962560 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963093 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0703df6-2aa8-4de5-b54e-850b63effded" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963116 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0703df6-2aa8-4de5-b54e-850b63effded" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963144 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="sg-core" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963152 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="sg-core" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963164 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aed630ee-8d81-4c05-8cd0-85245b99fa00" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963172 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="aed630ee-8d81-4c05-8cd0-85245b99fa00" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963196 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-central-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963204 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-central-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963219 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="proxy-httpd" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963228 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="proxy-httpd" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963251 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-notification-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963258 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-notification-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: E1009 08:36:48.963272 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="896adf7e-0f68-4f60-bd0c-6e507673c221" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963280 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="896adf7e-0f68-4f60-bd0c-6e507673c221" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963492 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="896adf7e-0f68-4f60-bd0c-6e507673c221" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963506 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-notification-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963526 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="sg-core" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963541 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0703df6-2aa8-4de5-b54e-850b63effded" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963552 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="ceilometer-central-agent" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963576 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" containerName="proxy-httpd" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.963589 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="aed630ee-8d81-4c05-8cd0-85245b99fa00" containerName="mariadb-database-create" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.965683 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.976116 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a75b743a-2d4c-4518-a829-71ae90478f45-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.976799 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.980561 4872 scope.go:117] "RemoveContainer" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.982392 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:36:48 crc kubenswrapper[4872]: I1009 08:36:48.982591 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.013234 4872 scope.go:117] "RemoveContainer" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:49 crc kubenswrapper[4872]: E1009 08:36:49.013696 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": container with ID starting with 1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9 not found: ID does not exist" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.013762 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} err="failed to get container status \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": rpc error: code = NotFound desc = could not find container \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": container with ID starting with 1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.013794 4872 scope.go:117] "RemoveContainer" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:49 crc kubenswrapper[4872]: E1009 08:36:49.014124 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": container with ID starting with 536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39 not found: ID does not exist" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014155 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} err="failed to get container status \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": rpc error: code = NotFound desc = could not find container \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": container with ID starting with 536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014175 4872 scope.go:117] "RemoveContainer" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:49 crc kubenswrapper[4872]: E1009 08:36:49.014328 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": container with ID starting with 91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f not found: ID does not exist" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014345 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} err="failed to get container status \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": rpc error: code = NotFound desc = could not find container \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": container with ID starting with 91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014358 4872 scope.go:117] "RemoveContainer" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:49 crc kubenswrapper[4872]: E1009 08:36:49.014763 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": container with ID starting with 0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344 not found: ID does not exist" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014795 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} err="failed to get container status \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": rpc error: code = NotFound desc = could not find container \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": container with ID starting with 0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.014836 4872 scope.go:117] "RemoveContainer" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015097 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} err="failed to get container status \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": rpc error: code = NotFound desc = could not find container \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": container with ID starting with 1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015117 4872 scope.go:117] "RemoveContainer" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015356 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} err="failed to get container status \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": rpc error: code = NotFound desc = could not find container \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": container with ID starting with 536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015401 4872 scope.go:117] "RemoveContainer" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015583 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} err="failed to get container status \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": rpc error: code = NotFound desc = could not find container \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": container with ID starting with 91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015609 4872 scope.go:117] "RemoveContainer" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015906 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} err="failed to get container status \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": rpc error: code = NotFound desc = could not find container \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": container with ID starting with 0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.015953 4872 scope.go:117] "RemoveContainer" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016140 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} err="failed to get container status \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": rpc error: code = NotFound desc = could not find container \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": container with ID starting with 1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016159 4872 scope.go:117] "RemoveContainer" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016296 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} err="failed to get container status \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": rpc error: code = NotFound desc = could not find container \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": container with ID starting with 536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016318 4872 scope.go:117] "RemoveContainer" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016499 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} err="failed to get container status \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": rpc error: code = NotFound desc = could not find container \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": container with ID starting with 91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016517 4872 scope.go:117] "RemoveContainer" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016698 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} err="failed to get container status \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": rpc error: code = NotFound desc = could not find container \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": container with ID starting with 0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.016722 4872 scope.go:117] "RemoveContainer" containerID="1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.017778 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9"} err="failed to get container status \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": rpc error: code = NotFound desc = could not find container \"1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9\": container with ID starting with 1ea5a1f954bc9e088a3a170396cce2b80ef09926b33f0a014faeecc80ffe53e9 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.017800 4872 scope.go:117] "RemoveContainer" containerID="536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.018005 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39"} err="failed to get container status \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": rpc error: code = NotFound desc = could not find container \"536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39\": container with ID starting with 536169320feb1fef30b26d2cd9f00620ca0e30cb36cb83567fbec30e9c30ce39 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.018023 4872 scope.go:117] "RemoveContainer" containerID="91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.018213 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f"} err="failed to get container status \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": rpc error: code = NotFound desc = could not find container \"91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f\": container with ID starting with 91f141039394689834ac228cc63892125aacc89c125ea94607ee78f61546802f not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.018231 4872 scope.go:117] "RemoveContainer" containerID="0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.018417 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344"} err="failed to get container status \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": rpc error: code = NotFound desc = could not find container \"0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344\": container with ID starting with 0d1242525731bc132feff2c31a9bdd74b3cc05d77459cb130bded75cc4e76344 not found: ID does not exist" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077500 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077595 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077628 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077665 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077703 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcdps\" (UniqueName: \"kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077788 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.077810 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179685 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179726 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179778 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcdps\" (UniqueName: \"kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179864 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179900 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.179933 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.180676 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.181089 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.184217 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.185307 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.186084 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.187256 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.200851 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcdps\" (UniqueName: \"kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps\") pod \"ceilometer-0\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.297605 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.830138 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:49 crc kubenswrapper[4872]: I1009 08:36:49.898131 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerStarted","Data":"1611121cfd4e772d5b0cae7bda84928ad81225e899d16ec1d6fbf99562b34b1f"} Oct 09 08:36:50 crc kubenswrapper[4872]: I1009 08:36:50.472918 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a75b743a-2d4c-4518-a829-71ae90478f45" path="/var/lib/kubelet/pods/a75b743a-2d4c-4518-a829-71ae90478f45/volumes" Oct 09 08:36:50 crc kubenswrapper[4872]: I1009 08:36:50.920568 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerStarted","Data":"40998c1db51640c173a27d1fad567a970b123646b9d86f190bdc82acc2117d5b"} Oct 09 08:36:51 crc kubenswrapper[4872]: I1009 08:36:51.264271 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:51 crc kubenswrapper[4872]: I1009 08:36:51.930515 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerStarted","Data":"89d4eebea5a8024622f21a51116bde8eae398b1f04d47d2a8a7f09aa71806e15"} Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.452738 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642258 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642336 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642373 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642431 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fc6n4\" (UniqueName: \"kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642508 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642566 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.642631 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle\") pod \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\" (UID: \"25fb62f7-d4a7-4973-8b0a-3857461a23d5\") " Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.647661 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs" (OuterVolumeSpecName: "logs") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.657061 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.670871 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4" (OuterVolumeSpecName: "kube-api-access-fc6n4") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "kube-api-access-fc6n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.685934 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data" (OuterVolumeSpecName: "config-data") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.720306 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts" (OuterVolumeSpecName: "scripts") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.723553 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.737423 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "25fb62f7-d4a7-4973-8b0a-3857461a23d5" (UID: "25fb62f7-d4a7-4973-8b0a-3857461a23d5"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744895 4872 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744933 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744947 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/25fb62f7-d4a7-4973-8b0a-3857461a23d5-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744958 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fc6n4\" (UniqueName: \"kubernetes.io/projected/25fb62f7-d4a7-4973-8b0a-3857461a23d5-kube-api-access-fc6n4\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744973 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/25fb62f7-d4a7-4973-8b0a-3857461a23d5-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744984 4872 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.744996 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25fb62f7-d4a7-4973-8b0a-3857461a23d5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.942836 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerStarted","Data":"3494162a3e3698cd8e31c925d739277c478add398694f6f72bb17efee26da9c7"} Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.944821 4872 generic.go:334] "Generic (PLEG): container finished" podID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerID="648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70" exitCode=137 Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.944869 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerDied","Data":"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70"} Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.944897 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cdd4b976-7cgml" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.944932 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cdd4b976-7cgml" event={"ID":"25fb62f7-d4a7-4973-8b0a-3857461a23d5","Type":"ContainerDied","Data":"782279b8f5b17fac627713657aea82da644534dcc851c1804e12f693d5fc92a2"} Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.944966 4872 scope.go:117] "RemoveContainer" containerID="61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7" Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.981256 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:36:52 crc kubenswrapper[4872]: I1009 08:36:52.987223 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6cdd4b976-7cgml"] Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.115449 4872 scope.go:117] "RemoveContainer" containerID="648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.148877 4872 scope.go:117] "RemoveContainer" containerID="61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7" Oct 09 08:36:53 crc kubenswrapper[4872]: E1009 08:36:53.149209 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7\": container with ID starting with 61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7 not found: ID does not exist" containerID="61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.149351 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7"} err="failed to get container status \"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7\": rpc error: code = NotFound desc = could not find container \"61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7\": container with ID starting with 61cd0fba912cd3b47377729e4b9a0befa2823c8facf4a4265bdaabce834297a7 not found: ID does not exist" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.150546 4872 scope.go:117] "RemoveContainer" containerID="648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70" Oct 09 08:36:53 crc kubenswrapper[4872]: E1009 08:36:53.150974 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70\": container with ID starting with 648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70 not found: ID does not exist" containerID="648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.151005 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70"} err="failed to get container status \"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70\": rpc error: code = NotFound desc = could not find container \"648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70\": container with ID starting with 648d54f95bc97b7cae277041842e723c688a76bc2e5990bcb5fd72a9bd93bd70 not found: ID does not exist" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.957575 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerStarted","Data":"c20dbe6431b70441087cefb5a02b1be25d3aba339f615868a0f8c9482a0e240d"} Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.957709 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-central-agent" containerID="cri-o://40998c1db51640c173a27d1fad567a970b123646b9d86f190bdc82acc2117d5b" gracePeriod=30 Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.958006 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.958060 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="proxy-httpd" containerID="cri-o://c20dbe6431b70441087cefb5a02b1be25d3aba339f615868a0f8c9482a0e240d" gracePeriod=30 Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.958071 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-notification-agent" containerID="cri-o://89d4eebea5a8024622f21a51116bde8eae398b1f04d47d2a8a7f09aa71806e15" gracePeriod=30 Oct 09 08:36:53 crc kubenswrapper[4872]: I1009 08:36:53.958156 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="sg-core" containerID="cri-o://3494162a3e3698cd8e31c925d739277c478add398694f6f72bb17efee26da9c7" gracePeriod=30 Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.202320 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.202366 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.231474 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.246318 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.255553 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.914054407 podStartE2EDuration="6.255532727s" podCreationTimestamp="2025-10-09 08:36:48 +0000 UTC" firstStartedPulling="2025-10-09 08:36:49.827578216 +0000 UTC m=+1168.018106842" lastFinishedPulling="2025-10-09 08:36:53.169056536 +0000 UTC m=+1171.359585162" observedRunningTime="2025-10-09 08:36:53.984967398 +0000 UTC m=+1172.175496034" watchObservedRunningTime="2025-10-09 08:36:54.255532727 +0000 UTC m=+1172.446061353" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.474428 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" path="/var/lib/kubelet/pods/25fb62f7-d4a7-4973-8b0a-3857461a23d5/volumes" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.976475 4872 generic.go:334] "Generic (PLEG): container finished" podID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerID="c20dbe6431b70441087cefb5a02b1be25d3aba339f615868a0f8c9482a0e240d" exitCode=0 Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.976802 4872 generic.go:334] "Generic (PLEG): container finished" podID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerID="3494162a3e3698cd8e31c925d739277c478add398694f6f72bb17efee26da9c7" exitCode=2 Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.976810 4872 generic.go:334] "Generic (PLEG): container finished" podID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerID="89d4eebea5a8024622f21a51116bde8eae398b1f04d47d2a8a7f09aa71806e15" exitCode=0 Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.976545 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerDied","Data":"c20dbe6431b70441087cefb5a02b1be25d3aba339f615868a0f8c9482a0e240d"} Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.978049 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.978068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerDied","Data":"3494162a3e3698cd8e31c925d739277c478add398694f6f72bb17efee26da9c7"} Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.978084 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 09 08:36:54 crc kubenswrapper[4872]: I1009 08:36:54.978095 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerDied","Data":"89d4eebea5a8024622f21a51116bde8eae398b1f04d47d2a8a7f09aa71806e15"} Oct 09 08:36:55 crc kubenswrapper[4872]: I1009 08:36:55.259378 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:55 crc kubenswrapper[4872]: I1009 08:36:55.259446 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:55 crc kubenswrapper[4872]: I1009 08:36:55.294228 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:55 crc kubenswrapper[4872]: I1009 08:36:55.302697 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:56 crc kubenswrapper[4872]: I1009 08:36:56.002900 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:56 crc kubenswrapper[4872]: I1009 08:36:56.002943 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.026353 4872 generic.go:334] "Generic (PLEG): container finished" podID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerID="40998c1db51640c173a27d1fad567a970b123646b9d86f190bdc82acc2117d5b" exitCode=0 Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.026739 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.026748 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.026400 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerDied","Data":"40998c1db51640c173a27d1fad567a970b123646b9d86f190bdc82acc2117d5b"} Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.241772 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.243468 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.334993 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.431905 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432068 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432158 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432232 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcdps\" (UniqueName: \"kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432266 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432315 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.432344 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data\") pod \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\" (UID: \"8460b9f6-a2bb-491a-8616-3db4cb6b9c58\") " Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.433051 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.433075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.437887 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps" (OuterVolumeSpecName: "kube-api-access-dcdps") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "kube-api-access-dcdps". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.438892 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts" (OuterVolumeSpecName: "scripts") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.470199 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.535050 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.535082 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.535092 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcdps\" (UniqueName: \"kubernetes.io/projected/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-kube-api-access-dcdps\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.535102 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.535110 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.536852 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.565548 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data" (OuterVolumeSpecName: "config-data") pod "8460b9f6-a2bb-491a-8616-3db4cb6b9c58" (UID: "8460b9f6-a2bb-491a-8616-3db4cb6b9c58"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.636692 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:57 crc kubenswrapper[4872]: I1009 08:36:57.636727 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8460b9f6-a2bb-491a-8616-3db4cb6b9c58-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.038444 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8460b9f6-a2bb-491a-8616-3db4cb6b9c58","Type":"ContainerDied","Data":"1611121cfd4e772d5b0cae7bda84928ad81225e899d16ec1d6fbf99562b34b1f"} Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.038514 4872 scope.go:117] "RemoveContainer" containerID="c20dbe6431b70441087cefb5a02b1be25d3aba339f615868a0f8c9482a0e240d" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.038530 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.038548 4872 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.038762 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.070706 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.074943 4872 scope.go:117] "RemoveContainer" containerID="3494162a3e3698cd8e31c925d739277c478add398694f6f72bb17efee26da9c7" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.078542 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097386 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097810 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="sg-core" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097833 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="sg-core" Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097844 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="proxy-httpd" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097850 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="proxy-httpd" Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097863 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-central-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097869 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-central-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097887 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097892 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097912 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon-log" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097918 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon-log" Oct 09 08:36:58 crc kubenswrapper[4872]: E1009 08:36:58.097936 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-notification-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.097945 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-notification-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098094 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="sg-core" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098113 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-notification-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098121 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon-log" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098132 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="25fb62f7-d4a7-4973-8b0a-3857461a23d5" containerName="horizon" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098139 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="proxy-httpd" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.098151 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" containerName="ceilometer-central-agent" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.101573 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.104537 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.112776 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.114035 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.117936 4872 scope.go:117] "RemoveContainer" containerID="89d4eebea5a8024622f21a51116bde8eae398b1f04d47d2a8a7f09aa71806e15" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145264 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl95v\" (UniqueName: \"kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145318 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145409 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145505 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145528 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145584 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.145687 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.170146 4872 scope.go:117] "RemoveContainer" containerID="40998c1db51640c173a27d1fad567a970b123646b9d86f190bdc82acc2117d5b" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247531 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247550 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247599 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247672 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl95v\" (UniqueName: \"kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.247690 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.248623 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.249225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.253118 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.253737 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.264304 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl95v\" (UniqueName: \"kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.266288 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.277512 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts\") pod \"ceilometer-0\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.279645 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.406238 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.456973 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.500552 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8460b9f6-a2bb-491a-8616-3db4cb6b9c58" path="/var/lib/kubelet/pods/8460b9f6-a2bb-491a-8616-3db4cb6b9c58/volumes" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.623381 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-1b9f-account-create-k5422"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.625168 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.629659 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.644836 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1b9f-account-create-k5422"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.653115 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcmp7\" (UniqueName: \"kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7\") pod \"nova-api-1b9f-account-create-k5422\" (UID: \"a78072f1-019e-478e-8736-c267e5604d4d\") " pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.755382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcmp7\" (UniqueName: \"kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7\") pod \"nova-api-1b9f-account-create-k5422\" (UID: \"a78072f1-019e-478e-8736-c267e5604d4d\") " pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.777905 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcmp7\" (UniqueName: \"kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7\") pod \"nova-api-1b9f-account-create-k5422\" (UID: \"a78072f1-019e-478e-8736-c267e5604d4d\") " pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.824591 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9786-account-create-dz7zk"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.827283 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.831222 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.835340 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9786-account-create-dz7zk"] Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.856820 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgzrp\" (UniqueName: \"kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp\") pod \"nova-cell0-9786-account-create-dz7zk\" (UID: \"0531d081-32b1-4166-9134-cfc92932b17f\") " pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.948228 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.960529 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgzrp\" (UniqueName: \"kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp\") pod \"nova-cell0-9786-account-create-dz7zk\" (UID: \"0531d081-32b1-4166-9134-cfc92932b17f\") " pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:36:58 crc kubenswrapper[4872]: I1009 08:36:58.982082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgzrp\" (UniqueName: \"kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp\") pod \"nova-cell0-9786-account-create-dz7zk\" (UID: \"0531d081-32b1-4166-9134-cfc92932b17f\") " pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.019291 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-051f-account-create-4mkf8"] Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.020803 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.027306 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.039356 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-051f-account-create-4mkf8"] Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.087704 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.107720 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.164129 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knwsq\" (UniqueName: \"kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq\") pod \"nova-cell1-051f-account-create-4mkf8\" (UID: \"e640feb9-e26f-45e3-9e6b-06be8678076e\") " pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.167937 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.266144 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knwsq\" (UniqueName: \"kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq\") pod \"nova-cell1-051f-account-create-4mkf8\" (UID: \"e640feb9-e26f-45e3-9e6b-06be8678076e\") " pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.287689 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knwsq\" (UniqueName: \"kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq\") pod \"nova-cell1-051f-account-create-4mkf8\" (UID: \"e640feb9-e26f-45e3-9e6b-06be8678076e\") " pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.381849 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.537810 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9786-account-create-dz7zk"] Oct 09 08:36:59 crc kubenswrapper[4872]: I1009 08:36:59.546240 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-1b9f-account-create-k5422"] Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.067685 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-051f-account-create-4mkf8"] Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.069165 4872 generic.go:334] "Generic (PLEG): container finished" podID="a78072f1-019e-478e-8736-c267e5604d4d" containerID="7a2c046245acf6e146feb206c3b9b6c773233f5af097f5f281683cc399a43f08" exitCode=0 Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.069627 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1b9f-account-create-k5422" event={"ID":"a78072f1-019e-478e-8736-c267e5604d4d","Type":"ContainerDied","Data":"7a2c046245acf6e146feb206c3b9b6c773233f5af097f5f281683cc399a43f08"} Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.069675 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1b9f-account-create-k5422" event={"ID":"a78072f1-019e-478e-8736-c267e5604d4d","Type":"ContainerStarted","Data":"dd7654f82f4e3df3883fdb83eac8c35dd409b837136d5325a27cb722cc66a254"} Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.074220 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerStarted","Data":"640e99bd08e78de76ac33133abd1ede3be963d9f85ad80c1fb5ee95e2beb4167"} Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.076314 4872 generic.go:334] "Generic (PLEG): container finished" podID="0531d081-32b1-4166-9134-cfc92932b17f" containerID="2ac02529dc578dc33ef21aca83ed8fbc82b9be1573c450ae5147e58eb5c37167" exitCode=0 Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.078122 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9786-account-create-dz7zk" event={"ID":"0531d081-32b1-4166-9134-cfc92932b17f","Type":"ContainerDied","Data":"2ac02529dc578dc33ef21aca83ed8fbc82b9be1573c450ae5147e58eb5c37167"} Oct 09 08:37:00 crc kubenswrapper[4872]: I1009 08:37:00.078181 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9786-account-create-dz7zk" event={"ID":"0531d081-32b1-4166-9134-cfc92932b17f","Type":"ContainerStarted","Data":"ebe65b0b07816e130291ab4afdcf9ad08146d8cfd7a911d0f18740f96b3459bd"} Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.091868 4872 generic.go:334] "Generic (PLEG): container finished" podID="e640feb9-e26f-45e3-9e6b-06be8678076e" containerID="823e451745a47d839ae0ddb052a09bbc3b7474a2b065da9f764e0bc14ae702da" exitCode=0 Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.093072 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-051f-account-create-4mkf8" event={"ID":"e640feb9-e26f-45e3-9e6b-06be8678076e","Type":"ContainerDied","Data":"823e451745a47d839ae0ddb052a09bbc3b7474a2b065da9f764e0bc14ae702da"} Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.093722 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-051f-account-create-4mkf8" event={"ID":"e640feb9-e26f-45e3-9e6b-06be8678076e","Type":"ContainerStarted","Data":"63e8f829bd5e7a69eaf4c34cf53f07e09d6c8c5325afe26f17b8ee5e39db1ca2"} Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.500203 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.516528 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.644531 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgzrp\" (UniqueName: \"kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp\") pod \"0531d081-32b1-4166-9134-cfc92932b17f\" (UID: \"0531d081-32b1-4166-9134-cfc92932b17f\") " Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.644612 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcmp7\" (UniqueName: \"kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7\") pod \"a78072f1-019e-478e-8736-c267e5604d4d\" (UID: \"a78072f1-019e-478e-8736-c267e5604d4d\") " Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.652715 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp" (OuterVolumeSpecName: "kube-api-access-fgzrp") pod "0531d081-32b1-4166-9134-cfc92932b17f" (UID: "0531d081-32b1-4166-9134-cfc92932b17f"). InnerVolumeSpecName "kube-api-access-fgzrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.652774 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7" (OuterVolumeSpecName: "kube-api-access-gcmp7") pod "a78072f1-019e-478e-8736-c267e5604d4d" (UID: "a78072f1-019e-478e-8736-c267e5604d4d"). InnerVolumeSpecName "kube-api-access-gcmp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.747267 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgzrp\" (UniqueName: \"kubernetes.io/projected/0531d081-32b1-4166-9134-cfc92932b17f-kube-api-access-fgzrp\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:01 crc kubenswrapper[4872]: I1009 08:37:01.747310 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcmp7\" (UniqueName: \"kubernetes.io/projected/a78072f1-019e-478e-8736-c267e5604d4d-kube-api-access-gcmp7\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.110046 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerStarted","Data":"c9124f5c427f0b6e9947877c86bc5ff8e36566d86fd7b622f44e27d0b770cbe2"} Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.111449 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerStarted","Data":"325258b8b0df55406ed3bed5fa1db0c64173bf0f39eb2b4fa9441e158acddb88"} Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.113583 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9786-account-create-dz7zk" event={"ID":"0531d081-32b1-4166-9134-cfc92932b17f","Type":"ContainerDied","Data":"ebe65b0b07816e130291ab4afdcf9ad08146d8cfd7a911d0f18740f96b3459bd"} Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.113753 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe65b0b07816e130291ab4afdcf9ad08146d8cfd7a911d0f18740f96b3459bd" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.113902 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9786-account-create-dz7zk" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.128042 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-1b9f-account-create-k5422" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.133884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-1b9f-account-create-k5422" event={"ID":"a78072f1-019e-478e-8736-c267e5604d4d","Type":"ContainerDied","Data":"dd7654f82f4e3df3883fdb83eac8c35dd409b837136d5325a27cb722cc66a254"} Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.133969 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd7654f82f4e3df3883fdb83eac8c35dd409b837136d5325a27cb722cc66a254" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.520046 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.662925 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-knwsq\" (UniqueName: \"kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq\") pod \"e640feb9-e26f-45e3-9e6b-06be8678076e\" (UID: \"e640feb9-e26f-45e3-9e6b-06be8678076e\") " Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.674993 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq" (OuterVolumeSpecName: "kube-api-access-knwsq") pod "e640feb9-e26f-45e3-9e6b-06be8678076e" (UID: "e640feb9-e26f-45e3-9e6b-06be8678076e"). InnerVolumeSpecName "kube-api-access-knwsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:02 crc kubenswrapper[4872]: I1009 08:37:02.766208 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-knwsq\" (UniqueName: \"kubernetes.io/projected/e640feb9-e26f-45e3-9e6b-06be8678076e-kube-api-access-knwsq\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:03 crc kubenswrapper[4872]: I1009 08:37:03.137633 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerStarted","Data":"495a63c50733271b73b6c75ca1693cb06dac9970845076c1560047ee3968c329"} Oct 09 08:37:03 crc kubenswrapper[4872]: I1009 08:37:03.140408 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-051f-account-create-4mkf8" event={"ID":"e640feb9-e26f-45e3-9e6b-06be8678076e","Type":"ContainerDied","Data":"63e8f829bd5e7a69eaf4c34cf53f07e09d6c8c5325afe26f17b8ee5e39db1ca2"} Oct 09 08:37:03 crc kubenswrapper[4872]: I1009 08:37:03.140532 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63e8f829bd5e7a69eaf4c34cf53f07e09d6c8c5325afe26f17b8ee5e39db1ca2" Oct 09 08:37:03 crc kubenswrapper[4872]: I1009 08:37:03.140652 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-051f-account-create-4mkf8" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.152913 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerStarted","Data":"08b5db50451904324b353632f0e158589373bed32bbca66d64af1061d2fa138e"} Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.153303 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.153131 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="sg-core" containerID="cri-o://495a63c50733271b73b6c75ca1693cb06dac9970845076c1560047ee3968c329" gracePeriod=30 Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.153037 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-central-agent" containerID="cri-o://325258b8b0df55406ed3bed5fa1db0c64173bf0f39eb2b4fa9441e158acddb88" gracePeriod=30 Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.153175 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-notification-agent" containerID="cri-o://c9124f5c427f0b6e9947877c86bc5ff8e36566d86fd7b622f44e27d0b770cbe2" gracePeriod=30 Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.153191 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="proxy-httpd" containerID="cri-o://08b5db50451904324b353632f0e158589373bed32bbca66d64af1061d2fa138e" gracePeriod=30 Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.181540 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.473909353 podStartE2EDuration="6.181522357s" podCreationTimestamp="2025-10-09 08:36:58 +0000 UTC" firstStartedPulling="2025-10-09 08:36:59.121180422 +0000 UTC m=+1177.311709048" lastFinishedPulling="2025-10-09 08:37:03.828793426 +0000 UTC m=+1182.019322052" observedRunningTime="2025-10-09 08:37:04.174500844 +0000 UTC m=+1182.365029470" watchObservedRunningTime="2025-10-09 08:37:04.181522357 +0000 UTC m=+1182.372050983" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.263616 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b7nfs"] Oct 09 08:37:04 crc kubenswrapper[4872]: E1009 08:37:04.264007 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a78072f1-019e-478e-8736-c267e5604d4d" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264023 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a78072f1-019e-478e-8736-c267e5604d4d" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: E1009 08:37:04.264056 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0531d081-32b1-4166-9134-cfc92932b17f" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264062 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0531d081-32b1-4166-9134-cfc92932b17f" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: E1009 08:37:04.264077 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e640feb9-e26f-45e3-9e6b-06be8678076e" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264083 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e640feb9-e26f-45e3-9e6b-06be8678076e" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264256 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e640feb9-e26f-45e3-9e6b-06be8678076e" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264281 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0531d081-32b1-4166-9134-cfc92932b17f" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264293 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a78072f1-019e-478e-8736-c267e5604d4d" containerName="mariadb-account-create" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.264878 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.267262 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.267351 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9gpmk" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.267585 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.279987 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b7nfs"] Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.402336 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.402742 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqdc9\" (UniqueName: \"kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.403110 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.403140 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.505020 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.505060 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.505110 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.505131 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqdc9\" (UniqueName: \"kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.512409 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.515431 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.521209 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.525696 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqdc9\" (UniqueName: \"kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9\") pod \"nova-cell0-conductor-db-sync-b7nfs\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:04 crc kubenswrapper[4872]: I1009 08:37:04.637081 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.136121 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b7nfs"] Oct 09 08:37:05 crc kubenswrapper[4872]: W1009 08:37:05.141873 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod968a20fd_dd60_49f3_8e98_0d7b9e059328.slice/crio-eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98 WatchSource:0}: Error finding container eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98: Status 404 returned error can't find the container with id eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98 Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.164355 4872 generic.go:334] "Generic (PLEG): container finished" podID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerID="495a63c50733271b73b6c75ca1693cb06dac9970845076c1560047ee3968c329" exitCode=2 Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.164392 4872 generic.go:334] "Generic (PLEG): container finished" podID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerID="c9124f5c427f0b6e9947877c86bc5ff8e36566d86fd7b622f44e27d0b770cbe2" exitCode=0 Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.164441 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerDied","Data":"495a63c50733271b73b6c75ca1693cb06dac9970845076c1560047ee3968c329"} Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.164467 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerDied","Data":"c9124f5c427f0b6e9947877c86bc5ff8e36566d86fd7b622f44e27d0b770cbe2"} Oct 09 08:37:05 crc kubenswrapper[4872]: I1009 08:37:05.165709 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" event={"ID":"968a20fd-dd60-49f3-8e98-0d7b9e059328","Type":"ContainerStarted","Data":"eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98"} Oct 09 08:37:09 crc kubenswrapper[4872]: I1009 08:37:09.205005 4872 generic.go:334] "Generic (PLEG): container finished" podID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerID="325258b8b0df55406ed3bed5fa1db0c64173bf0f39eb2b4fa9441e158acddb88" exitCode=0 Oct 09 08:37:09 crc kubenswrapper[4872]: I1009 08:37:09.205042 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerDied","Data":"325258b8b0df55406ed3bed5fa1db0c64173bf0f39eb2b4fa9441e158acddb88"} Oct 09 08:37:15 crc kubenswrapper[4872]: I1009 08:37:15.273871 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" event={"ID":"968a20fd-dd60-49f3-8e98-0d7b9e059328","Type":"ContainerStarted","Data":"9b57439369b1d66ba5273fd53bf568d3cdc851b60932530e41a5606ac41aed04"} Oct 09 08:37:15 crc kubenswrapper[4872]: I1009 08:37:15.291048 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" podStartSLOduration=2.174670559 podStartE2EDuration="11.2910317s" podCreationTimestamp="2025-10-09 08:37:04 +0000 UTC" firstStartedPulling="2025-10-09 08:37:05.144827663 +0000 UTC m=+1183.335356289" lastFinishedPulling="2025-10-09 08:37:14.261188804 +0000 UTC m=+1192.451717430" observedRunningTime="2025-10-09 08:37:15.288086975 +0000 UTC m=+1193.478615611" watchObservedRunningTime="2025-10-09 08:37:15.2910317 +0000 UTC m=+1193.481560326" Oct 09 08:37:26 crc kubenswrapper[4872]: I1009 08:37:26.409655 4872 generic.go:334] "Generic (PLEG): container finished" podID="968a20fd-dd60-49f3-8e98-0d7b9e059328" containerID="9b57439369b1d66ba5273fd53bf568d3cdc851b60932530e41a5606ac41aed04" exitCode=0 Oct 09 08:37:26 crc kubenswrapper[4872]: I1009 08:37:26.409791 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" event={"ID":"968a20fd-dd60-49f3-8e98-0d7b9e059328","Type":"ContainerDied","Data":"9b57439369b1d66ba5273fd53bf568d3cdc851b60932530e41a5606ac41aed04"} Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.834828 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.941266 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle\") pod \"968a20fd-dd60-49f3-8e98-0d7b9e059328\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.941366 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqdc9\" (UniqueName: \"kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9\") pod \"968a20fd-dd60-49f3-8e98-0d7b9e059328\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.941451 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts\") pod \"968a20fd-dd60-49f3-8e98-0d7b9e059328\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.941484 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data\") pod \"968a20fd-dd60-49f3-8e98-0d7b9e059328\" (UID: \"968a20fd-dd60-49f3-8e98-0d7b9e059328\") " Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.948180 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9" (OuterVolumeSpecName: "kube-api-access-bqdc9") pod "968a20fd-dd60-49f3-8e98-0d7b9e059328" (UID: "968a20fd-dd60-49f3-8e98-0d7b9e059328"). InnerVolumeSpecName "kube-api-access-bqdc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.948425 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts" (OuterVolumeSpecName: "scripts") pod "968a20fd-dd60-49f3-8e98-0d7b9e059328" (UID: "968a20fd-dd60-49f3-8e98-0d7b9e059328"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.969731 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data" (OuterVolumeSpecName: "config-data") pod "968a20fd-dd60-49f3-8e98-0d7b9e059328" (UID: "968a20fd-dd60-49f3-8e98-0d7b9e059328"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:27 crc kubenswrapper[4872]: I1009 08:37:27.978111 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "968a20fd-dd60-49f3-8e98-0d7b9e059328" (UID: "968a20fd-dd60-49f3-8e98-0d7b9e059328"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.043762 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.043802 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqdc9\" (UniqueName: \"kubernetes.io/projected/968a20fd-dd60-49f3-8e98-0d7b9e059328-kube-api-access-bqdc9\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.043820 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.043833 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/968a20fd-dd60-49f3-8e98-0d7b9e059328-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.429762 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" event={"ID":"968a20fd-dd60-49f3-8e98-0d7b9e059328","Type":"ContainerDied","Data":"eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98"} Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.429802 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb8d1b724d540ccdc73d9ae0ffdb3bdf944265eb38efcea6550e18d4f8ec2a98" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.429850 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b7nfs" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.468448 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.553611 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 08:37:28 crc kubenswrapper[4872]: E1009 08:37:28.554288 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="968a20fd-dd60-49f3-8e98-0d7b9e059328" containerName="nova-cell0-conductor-db-sync" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.554316 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="968a20fd-dd60-49f3-8e98-0d7b9e059328" containerName="nova-cell0-conductor-db-sync" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.554484 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="968a20fd-dd60-49f3-8e98-0d7b9e059328" containerName="nova-cell0-conductor-db-sync" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.555090 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.564454 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.591859 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-9gpmk" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.592137 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.655182 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mq8v\" (UniqueName: \"kubernetes.io/projected/c572e5bd-51c3-49b9-9c55-7e9f888926a2-kube-api-access-9mq8v\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.655329 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.655390 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.757228 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.757340 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mq8v\" (UniqueName: \"kubernetes.io/projected/c572e5bd-51c3-49b9-9c55-7e9f888926a2-kube-api-access-9mq8v\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.757408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.761876 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.762456 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c572e5bd-51c3-49b9-9c55-7e9f888926a2-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.775170 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mq8v\" (UniqueName: \"kubernetes.io/projected/c572e5bd-51c3-49b9-9c55-7e9f888926a2-kube-api-access-9mq8v\") pod \"nova-cell0-conductor-0\" (UID: \"c572e5bd-51c3-49b9-9c55-7e9f888926a2\") " pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:28 crc kubenswrapper[4872]: I1009 08:37:28.921087 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:29 crc kubenswrapper[4872]: I1009 08:37:29.385300 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 09 08:37:29 crc kubenswrapper[4872]: I1009 08:37:29.440181 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c572e5bd-51c3-49b9-9c55-7e9f888926a2","Type":"ContainerStarted","Data":"cb36794be28cb4f4986e96ce66bc2b73d8fd99949c4107c8c5047e01e6a1805d"} Oct 09 08:37:30 crc kubenswrapper[4872]: I1009 08:37:30.451058 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"c572e5bd-51c3-49b9-9c55-7e9f888926a2","Type":"ContainerStarted","Data":"fd0ecab4e47de018981e1c5196453d6351dbaeb44fa6534ddd9327e7887fdc08"} Oct 09 08:37:30 crc kubenswrapper[4872]: I1009 08:37:30.451772 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:30 crc kubenswrapper[4872]: I1009 08:37:30.474576 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.474556077 podStartE2EDuration="2.474556077s" podCreationTimestamp="2025-10-09 08:37:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:30.468758039 +0000 UTC m=+1208.659286675" watchObservedRunningTime="2025-10-09 08:37:30.474556077 +0000 UTC m=+1208.665084723" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.496095 4872 generic.go:334] "Generic (PLEG): container finished" podID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerID="08b5db50451904324b353632f0e158589373bed32bbca66d64af1061d2fa138e" exitCode=137 Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.496182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerDied","Data":"08b5db50451904324b353632f0e158589373bed32bbca66d64af1061d2fa138e"} Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.608301 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773079 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773143 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773187 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl95v\" (UniqueName: \"kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773348 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773381 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773406 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773465 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts\") pod \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\" (UID: \"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661\") " Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.773937 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.774548 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.779012 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts" (OuterVolumeSpecName: "scripts") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.779024 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v" (OuterVolumeSpecName: "kube-api-access-xl95v") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "kube-api-access-xl95v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.800518 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.852855 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.867985 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data" (OuterVolumeSpecName: "config-data") pod "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" (UID: "2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875892 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875920 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875930 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl95v\" (UniqueName: \"kubernetes.io/projected/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-kube-api-access-xl95v\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875941 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875951 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875959 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:34 crc kubenswrapper[4872]: I1009 08:37:34.875966 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.543687 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661","Type":"ContainerDied","Data":"640e99bd08e78de76ac33133abd1ede3be963d9f85ad80c1fb5ee95e2beb4167"} Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.543744 4872 scope.go:117] "RemoveContainer" containerID="08b5db50451904324b353632f0e158589373bed32bbca66d64af1061d2fa138e" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.543762 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.565297 4872 scope.go:117] "RemoveContainer" containerID="495a63c50733271b73b6c75ca1693cb06dac9970845076c1560047ee3968c329" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.578486 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.586896 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.613269 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:37:35 crc kubenswrapper[4872]: E1009 08:37:35.614165 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-notification-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614190 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-notification-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: E1009 08:37:35.614233 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-central-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614241 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-central-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: E1009 08:37:35.614254 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="sg-core" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614261 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="sg-core" Oct 09 08:37:35 crc kubenswrapper[4872]: E1009 08:37:35.614274 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="proxy-httpd" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614282 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="proxy-httpd" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614533 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="sg-core" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614547 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="proxy-httpd" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614566 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-notification-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.614581 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" containerName="ceilometer-central-agent" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.617821 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.618338 4872 scope.go:117] "RemoveContainer" containerID="c9124f5c427f0b6e9947877c86bc5ff8e36566d86fd7b622f44e27d0b770cbe2" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.620577 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.621480 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.650661 4872 scope.go:117] "RemoveContainer" containerID="325258b8b0df55406ed3bed5fa1db0c64173bf0f39eb2b4fa9441e158acddb88" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.664976 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.687858 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.687952 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.688121 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.688270 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdb2n\" (UniqueName: \"kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.688300 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.688320 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.688442 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789738 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789804 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789849 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdb2n\" (UniqueName: \"kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789868 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789885 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789929 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.789953 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.790474 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.790475 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.795211 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.795303 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.795583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.804828 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.805311 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdb2n\" (UniqueName: \"kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n\") pod \"ceilometer-0\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " pod="openstack/ceilometer-0" Oct 09 08:37:35 crc kubenswrapper[4872]: I1009 08:37:35.950048 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:37:36 crc kubenswrapper[4872]: W1009 08:37:36.413096 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14114366_3880_4288_9d9c_ffc0d7610046.slice/crio-db2e9f4e76c5cf40fc4d78c753687b8d2f3adfb3cdf74962c6f1aa84693aa196 WatchSource:0}: Error finding container db2e9f4e76c5cf40fc4d78c753687b8d2f3adfb3cdf74962c6f1aa84693aa196: Status 404 returned error can't find the container with id db2e9f4e76c5cf40fc4d78c753687b8d2f3adfb3cdf74962c6f1aa84693aa196 Oct 09 08:37:36 crc kubenswrapper[4872]: I1009 08:37:36.416210 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:37:36 crc kubenswrapper[4872]: I1009 08:37:36.489736 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661" path="/var/lib/kubelet/pods/2e4dcd9d-493f-4d8e-a4c0-e2835c5d1661/volumes" Oct 09 08:37:36 crc kubenswrapper[4872]: I1009 08:37:36.553787 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerStarted","Data":"db2e9f4e76c5cf40fc4d78c753687b8d2f3adfb3cdf74962c6f1aa84693aa196"} Oct 09 08:37:37 crc kubenswrapper[4872]: I1009 08:37:37.568894 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerStarted","Data":"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41"} Oct 09 08:37:38 crc kubenswrapper[4872]: I1009 08:37:38.586949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerStarted","Data":"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0"} Oct 09 08:37:38 crc kubenswrapper[4872]: I1009 08:37:38.951026 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.430203 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-htgks"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.436253 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.439486 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.442743 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-htgks"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.445898 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.593119 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.594701 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.596446 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.596498 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.596534 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8dvt\" (UniqueName: \"kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.596622 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.597488 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.600436 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.604677 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerStarted","Data":"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799"} Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.659447 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.661685 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.671382 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.682197 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.699585 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.699899 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.699942 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.699990 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8dvt\" (UniqueName: \"kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.700138 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.700163 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wlb7\" (UniqueName: \"kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.700203 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.710521 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.711110 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.731175 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.747275 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8dvt\" (UniqueName: \"kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt\") pod \"nova-cell0-cell-mapping-htgks\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.758407 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.799948 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.801485 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.809628 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812246 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wlb7\" (UniqueName: \"kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812299 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812326 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812348 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812476 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.817653 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.820476 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.820965 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.812499 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.823190 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv88k\" (UniqueName: \"kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.851362 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wlb7\" (UniqueName: \"kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7\") pod \"nova-cell1-novncproxy-0\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.871743 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.875098 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.877336 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.882604 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.896231 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.897887 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925237 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925354 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925483 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925510 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq5bf\" (UniqueName: \"kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925578 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.925605 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv88k\" (UniqueName: \"kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.927067 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.928830 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.933320 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.933648 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.941232 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.948257 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv88k\" (UniqueName: \"kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k\") pod \"nova-metadata-0\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " pod="openstack/nova-metadata-0" Oct 09 08:37:39 crc kubenswrapper[4872]: I1009 08:37:39.988373 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030181 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030249 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l85wj\" (UniqueName: \"kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030293 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030314 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030396 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030432 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq5bf\" (UniqueName: \"kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030455 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxmqb\" (UniqueName: \"kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030479 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030501 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030529 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030558 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.030578 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.047357 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.051805 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq5bf\" (UniqueName: \"kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.052527 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133031 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133394 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l85wj\" (UniqueName: \"kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133427 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133447 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133523 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxmqb\" (UniqueName: \"kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133540 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133556 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133574 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.133600 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.134072 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.134993 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.138631 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.141141 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.141783 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.142970 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.143037 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.145464 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.167312 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l85wj\" (UniqueName: \"kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj\") pod \"dnsmasq-dns-865f5d856f-c2c79\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.174077 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxmqb\" (UniqueName: \"kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb\") pod \"nova-api-0\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.299964 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.343841 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.355148 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.405237 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-htgks"] Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.573481 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.645044 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.649160 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerStarted","Data":"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53"} Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.649368 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.652065 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2336b0ba-f01d-4013-a125-dbfc3a5c384d","Type":"ContainerStarted","Data":"7945151881e59ff81aa5dc335fb5e86ef72147f1528b7b8a0445382688e130b5"} Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.656734 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-htgks" event={"ID":"96070559-6af7-4827-8dd7-38edfdd2ac47","Type":"ContainerStarted","Data":"f5dc3516e87821cd7e91d6174e5ca8ea40018ae25d0ac808104f3f2ca8f74850"} Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.678674 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.269641673 podStartE2EDuration="5.678656514s" podCreationTimestamp="2025-10-09 08:37:35 +0000 UTC" firstStartedPulling="2025-10-09 08:37:36.417793145 +0000 UTC m=+1214.608321781" lastFinishedPulling="2025-10-09 08:37:39.826808006 +0000 UTC m=+1218.017336622" observedRunningTime="2025-10-09 08:37:40.675831653 +0000 UTC m=+1218.866360309" watchObservedRunningTime="2025-10-09 08:37:40.678656514 +0000 UTC m=+1218.869185140" Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.915319 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:40 crc kubenswrapper[4872]: I1009 08:37:40.934228 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.009990 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hrvkd"] Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.011394 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.015141 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.015603 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.020144 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hrvkd"] Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.078077 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.164012 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.165836 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvl4b\" (UniqueName: \"kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.165952 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.166033 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.268083 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.268157 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvl4b\" (UniqueName: \"kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.268206 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.268235 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.278883 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.284130 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.292507 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.299516 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvl4b\" (UniqueName: \"kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b\") pod \"nova-cell1-conductor-db-sync-hrvkd\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.341335 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.676937 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ad5a-5664-4bc7-8847-146479f7d036","Type":"ContainerStarted","Data":"9ee1ca6a11e32098e8b2a8004ec3318f32f6406df03e52282f1c7d42257f6761"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.683889 4872 generic.go:334] "Generic (PLEG): container finished" podID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerID="f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0" exitCode=0 Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.684185 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" event={"ID":"a2e37277-59e1-49b6-a240-660d3416ff4d","Type":"ContainerDied","Data":"f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.684245 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" event={"ID":"a2e37277-59e1-49b6-a240-660d3416ff4d","Type":"ContainerStarted","Data":"5ee87ff791607adf76f8cbf99040e6130efbf537b343c599834da1d59738aace"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.688071 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-htgks" event={"ID":"96070559-6af7-4827-8dd7-38edfdd2ac47","Type":"ContainerStarted","Data":"73ce3c85dfcd9f73f5e0466e57db9103d94dfe90e36711c0083e2d6abdf2b25f"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.698275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerStarted","Data":"1e8851aae7890ea012c9b9d6960000c0d8b9a6f1ccdd380ed9f9d529356cd07d"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.710588 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerStarted","Data":"8ec79e351987ba59012bb49ae931df6b75687db7e22a43b8637184933a9f232e"} Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.746405 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-htgks" podStartSLOduration=2.746389123 podStartE2EDuration="2.746389123s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:41.737945479 +0000 UTC m=+1219.928474105" watchObservedRunningTime="2025-10-09 08:37:41.746389123 +0000 UTC m=+1219.936917749" Oct 09 08:37:41 crc kubenswrapper[4872]: I1009 08:37:41.789146 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hrvkd"] Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.722745 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" event={"ID":"3b76157d-7dc5-4661-972e-6673f3bd9990","Type":"ContainerStarted","Data":"a8f8e67dd4074a3a0cfa9f4f8b289b0f734ae072d14085b0b850339887d460bc"} Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.723839 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" event={"ID":"3b76157d-7dc5-4661-972e-6673f3bd9990","Type":"ContainerStarted","Data":"7fba23fb5bb0b8f7c4067e76f141c606a0f6c9f458cd00ba9edec12663017511"} Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.729445 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" event={"ID":"a2e37277-59e1-49b6-a240-660d3416ff4d","Type":"ContainerStarted","Data":"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2"} Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.729501 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.741120 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" podStartSLOduration=2.741099546 podStartE2EDuration="2.741099546s" podCreationTimestamp="2025-10-09 08:37:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:42.735402681 +0000 UTC m=+1220.925931307" watchObservedRunningTime="2025-10-09 08:37:42.741099546 +0000 UTC m=+1220.931628172" Oct 09 08:37:42 crc kubenswrapper[4872]: I1009 08:37:42.759975 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" podStartSLOduration=3.7599569600000002 podStartE2EDuration="3.75995696s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:42.755489131 +0000 UTC m=+1220.946017777" watchObservedRunningTime="2025-10-09 08:37:42.75995696 +0000 UTC m=+1220.950485586" Oct 09 08:37:43 crc kubenswrapper[4872]: I1009 08:37:43.752559 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:43 crc kubenswrapper[4872]: I1009 08:37:43.764228 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.779002 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerStarted","Data":"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.779586 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerStarted","Data":"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.779117 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-log" containerID="cri-o://00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" gracePeriod=30 Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.779608 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-metadata" containerID="cri-o://564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" gracePeriod=30 Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.789763 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerStarted","Data":"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.789809 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerStarted","Data":"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.807232 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.748496614 podStartE2EDuration="6.807215098s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="2025-10-09 08:37:40.644813887 +0000 UTC m=+1218.835342503" lastFinishedPulling="2025-10-09 08:37:44.703532361 +0000 UTC m=+1222.894060987" observedRunningTime="2025-10-09 08:37:45.795125409 +0000 UTC m=+1223.985654035" watchObservedRunningTime="2025-10-09 08:37:45.807215098 +0000 UTC m=+1223.997743724" Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.808107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ad5a-5664-4bc7-8847-146479f7d036","Type":"ContainerStarted","Data":"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.813062 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2336b0ba-f01d-4013-a125-dbfc3a5c384d","Type":"ContainerStarted","Data":"e5456a0bc88528481d0e92e61aeb1fa5c43395c2ec951ebab383d1b09cac6bf3"} Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.813219 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://e5456a0bc88528481d0e92e61aeb1fa5c43395c2ec951ebab383d1b09cac6bf3" gracePeriod=30 Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.842417 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.070595391 podStartE2EDuration="6.842392083s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="2025-10-09 08:37:40.931666167 +0000 UTC m=+1219.122194793" lastFinishedPulling="2025-10-09 08:37:44.703462859 +0000 UTC m=+1222.893991485" observedRunningTime="2025-10-09 08:37:45.818232576 +0000 UTC m=+1224.008761202" watchObservedRunningTime="2025-10-09 08:37:45.842392083 +0000 UTC m=+1224.032920729" Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.860470 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.753941031 podStartE2EDuration="6.860445144s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="2025-10-09 08:37:40.593210718 +0000 UTC m=+1218.783739344" lastFinishedPulling="2025-10-09 08:37:44.699714831 +0000 UTC m=+1222.890243457" observedRunningTime="2025-10-09 08:37:45.834968329 +0000 UTC m=+1224.025496955" watchObservedRunningTime="2025-10-09 08:37:45.860445144 +0000 UTC m=+1224.050973770" Oct 09 08:37:45 crc kubenswrapper[4872]: I1009 08:37:45.869358 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.083880135 podStartE2EDuration="6.869340591s" podCreationTimestamp="2025-10-09 08:37:39 +0000 UTC" firstStartedPulling="2025-10-09 08:37:40.917685654 +0000 UTC m=+1219.108214280" lastFinishedPulling="2025-10-09 08:37:44.70314611 +0000 UTC m=+1222.893674736" observedRunningTime="2025-10-09 08:37:45.849517689 +0000 UTC m=+1224.040046325" watchObservedRunningTime="2025-10-09 08:37:45.869340591 +0000 UTC m=+1224.059869217" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.406805 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.488893 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv88k\" (UniqueName: \"kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k\") pod \"d2772932-060d-4da6-83ec-464ef58511cc\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.502101 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle\") pod \"d2772932-060d-4da6-83ec-464ef58511cc\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.502197 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data\") pod \"d2772932-060d-4da6-83ec-464ef58511cc\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.502255 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs\") pod \"d2772932-060d-4da6-83ec-464ef58511cc\" (UID: \"d2772932-060d-4da6-83ec-464ef58511cc\") " Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.503296 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs" (OuterVolumeSpecName: "logs") pod "d2772932-060d-4da6-83ec-464ef58511cc" (UID: "d2772932-060d-4da6-83ec-464ef58511cc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.503660 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d2772932-060d-4da6-83ec-464ef58511cc-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.511912 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k" (OuterVolumeSpecName: "kube-api-access-tv88k") pod "d2772932-060d-4da6-83ec-464ef58511cc" (UID: "d2772932-060d-4da6-83ec-464ef58511cc"). InnerVolumeSpecName "kube-api-access-tv88k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.533407 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d2772932-060d-4da6-83ec-464ef58511cc" (UID: "d2772932-060d-4da6-83ec-464ef58511cc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.533818 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data" (OuterVolumeSpecName: "config-data") pod "d2772932-060d-4da6-83ec-464ef58511cc" (UID: "d2772932-060d-4da6-83ec-464ef58511cc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.605787 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv88k\" (UniqueName: \"kubernetes.io/projected/d2772932-060d-4da6-83ec-464ef58511cc-kube-api-access-tv88k\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.605823 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.605835 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2772932-060d-4da6-83ec-464ef58511cc-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.829775 4872 generic.go:334] "Generic (PLEG): container finished" podID="d2772932-060d-4da6-83ec-464ef58511cc" containerID="564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" exitCode=0 Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.830988 4872 generic.go:334] "Generic (PLEG): container finished" podID="d2772932-060d-4da6-83ec-464ef58511cc" containerID="00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" exitCode=143 Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.829949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerDied","Data":"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b"} Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.832197 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerDied","Data":"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f"} Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.832285 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d2772932-060d-4da6-83ec-464ef58511cc","Type":"ContainerDied","Data":"1e8851aae7890ea012c9b9d6960000c0d8b9a6f1ccdd380ed9f9d529356cd07d"} Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.832382 4872 scope.go:117] "RemoveContainer" containerID="564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.830040 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.876201 4872 scope.go:117] "RemoveContainer" containerID="00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.880794 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.902799 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.906836 4872 scope.go:117] "RemoveContainer" containerID="564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" Oct 09 08:37:46 crc kubenswrapper[4872]: E1009 08:37:46.907300 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b\": container with ID starting with 564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b not found: ID does not exist" containerID="564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.907352 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b"} err="failed to get container status \"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b\": rpc error: code = NotFound desc = could not find container \"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b\": container with ID starting with 564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b not found: ID does not exist" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.907374 4872 scope.go:117] "RemoveContainer" containerID="00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" Oct 09 08:37:46 crc kubenswrapper[4872]: E1009 08:37:46.907783 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f\": container with ID starting with 00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f not found: ID does not exist" containerID="00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.907876 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f"} err="failed to get container status \"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f\": rpc error: code = NotFound desc = could not find container \"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f\": container with ID starting with 00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f not found: ID does not exist" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.907952 4872 scope.go:117] "RemoveContainer" containerID="564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.908279 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b"} err="failed to get container status \"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b\": rpc error: code = NotFound desc = could not find container \"564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b\": container with ID starting with 564a26ebbdf7a87519fdb8c670d4af08afcb7754b3ec8c2ffe5b560549f99e1b not found: ID does not exist" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.908356 4872 scope.go:117] "RemoveContainer" containerID="00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.908666 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f"} err="failed to get container status \"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f\": rpc error: code = NotFound desc = could not find container \"00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f\": container with ID starting with 00ffb21c7672ec6f2f05c45c8be5b259a1bba2b105ec28de062dd337c7d5275f not found: ID does not exist" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.914540 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:46 crc kubenswrapper[4872]: E1009 08:37:46.915313 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-log" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.915384 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-log" Oct 09 08:37:46 crc kubenswrapper[4872]: E1009 08:37:46.915431 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-metadata" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.915451 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-metadata" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.915949 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-metadata" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.916036 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2772932-060d-4da6-83ec-464ef58511cc" containerName="nova-metadata-log" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.918479 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.922570 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.923835 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 08:37:46 crc kubenswrapper[4872]: I1009 08:37:46.924443 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.013692 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tb5q2\" (UniqueName: \"kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.013780 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.013902 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.013929 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.013979 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.115950 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.116011 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.116063 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.116134 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tb5q2\" (UniqueName: \"kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.116186 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.120595 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.126295 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.139011 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.139281 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.145761 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tb5q2\" (UniqueName: \"kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2\") pod \"nova-metadata-0\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.249463 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.770590 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:47 crc kubenswrapper[4872]: I1009 08:37:47.845126 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerStarted","Data":"220b66ffd75e98c7dc8cfdadcc7dbc7564acf82b32c345fe9c3b1d39fbbbffd0"} Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.473294 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2772932-060d-4da6-83ec-464ef58511cc" path="/var/lib/kubelet/pods/d2772932-060d-4da6-83ec-464ef58511cc/volumes" Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.856527 4872 generic.go:334] "Generic (PLEG): container finished" podID="96070559-6af7-4827-8dd7-38edfdd2ac47" containerID="73ce3c85dfcd9f73f5e0466e57db9103d94dfe90e36711c0083e2d6abdf2b25f" exitCode=0 Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.856633 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-htgks" event={"ID":"96070559-6af7-4827-8dd7-38edfdd2ac47","Type":"ContainerDied","Data":"73ce3c85dfcd9f73f5e0466e57db9103d94dfe90e36711c0083e2d6abdf2b25f"} Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.860658 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerStarted","Data":"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2"} Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.860696 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerStarted","Data":"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b"} Oct 09 08:37:48 crc kubenswrapper[4872]: I1009 08:37:48.908620 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.908596088 podStartE2EDuration="2.908596088s" podCreationTimestamp="2025-10-09 08:37:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:48.898772674 +0000 UTC m=+1227.089301310" watchObservedRunningTime="2025-10-09 08:37:48.908596088 +0000 UTC m=+1227.099124714" Oct 09 08:37:49 crc kubenswrapper[4872]: I1009 08:37:49.942348 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.259243 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.301127 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.301367 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.344587 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.344657 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.345042 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.357964 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.379709 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts\") pod \"96070559-6af7-4827-8dd7-38edfdd2ac47\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.380510 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data\") pod \"96070559-6af7-4827-8dd7-38edfdd2ac47\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.380886 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle\") pod \"96070559-6af7-4827-8dd7-38edfdd2ac47\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.381108 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8dvt\" (UniqueName: \"kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt\") pod \"96070559-6af7-4827-8dd7-38edfdd2ac47\" (UID: \"96070559-6af7-4827-8dd7-38edfdd2ac47\") " Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.402410 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt" (OuterVolumeSpecName: "kube-api-access-z8dvt") pod "96070559-6af7-4827-8dd7-38edfdd2ac47" (UID: "96070559-6af7-4827-8dd7-38edfdd2ac47"). InnerVolumeSpecName "kube-api-access-z8dvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.417889 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts" (OuterVolumeSpecName: "scripts") pod "96070559-6af7-4827-8dd7-38edfdd2ac47" (UID: "96070559-6af7-4827-8dd7-38edfdd2ac47"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.461849 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "96070559-6af7-4827-8dd7-38edfdd2ac47" (UID: "96070559-6af7-4827-8dd7-38edfdd2ac47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.496779 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data" (OuterVolumeSpecName: "config-data") pod "96070559-6af7-4827-8dd7-38edfdd2ac47" (UID: "96070559-6af7-4827-8dd7-38edfdd2ac47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.496889 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8dvt\" (UniqueName: \"kubernetes.io/projected/96070559-6af7-4827-8dd7-38edfdd2ac47-kube-api-access-z8dvt\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.496920 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.496932 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.511385 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.511654 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="dnsmasq-dns" containerID="cri-o://dfe5352f1ca2fd7daae4cceb44cada175049f883c7a926b8c9de8f52508e12c6" gracePeriod=10 Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.598373 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/96070559-6af7-4827-8dd7-38edfdd2ac47-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.899540 4872 generic.go:334] "Generic (PLEG): container finished" podID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerID="dfe5352f1ca2fd7daae4cceb44cada175049f883c7a926b8c9de8f52508e12c6" exitCode=0 Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.899656 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerDied","Data":"dfe5352f1ca2fd7daae4cceb44cada175049f883c7a926b8c9de8f52508e12c6"} Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.914230 4872 generic.go:334] "Generic (PLEG): container finished" podID="3b76157d-7dc5-4661-972e-6673f3bd9990" containerID="a8f8e67dd4074a3a0cfa9f4f8b289b0f734ae072d14085b0b850339887d460bc" exitCode=0 Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.914290 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" event={"ID":"3b76157d-7dc5-4661-972e-6673f3bd9990","Type":"ContainerDied","Data":"a8f8e67dd4074a3a0cfa9f4f8b289b0f734ae072d14085b0b850339887d460bc"} Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.917287 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-htgks" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.917712 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-htgks" event={"ID":"96070559-6af7-4827-8dd7-38edfdd2ac47","Type":"ContainerDied","Data":"f5dc3516e87821cd7e91d6174e5ca8ea40018ae25d0ac808104f3f2ca8f74850"} Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.917738 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5dc3516e87821cd7e91d6174e5ca8ea40018ae25d0ac808104f3f2ca8f74850" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.951239 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 08:37:50 crc kubenswrapper[4872]: I1009 08:37:50.957471 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.017407 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.017938 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pchsv\" (UniqueName: \"kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.018017 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.018113 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.018145 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.018195 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0\") pod \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\" (UID: \"3b4436f4-a95e-4b70-9e13-02f70d26ee6e\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.047909 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv" (OuterVolumeSpecName: "kube-api-access-pchsv") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "kube-api-access-pchsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.101259 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config" (OuterVolumeSpecName: "config") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.111538 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.112946 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.121871 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.121900 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.121912 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pchsv\" (UniqueName: \"kubernetes.io/projected/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-kube-api-access-pchsv\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.121920 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.126171 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.141115 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.142253 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3b4436f4-a95e-4b70-9e13-02f70d26ee6e" (UID: "3b4436f4-a95e-4b70-9e13-02f70d26ee6e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.147311 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.147619 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-log" containerID="cri-o://ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" gracePeriod=30 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.147909 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-metadata" containerID="cri-o://a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" gracePeriod=30 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.223062 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.223098 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3b4436f4-a95e-4b70-9e13-02f70d26ee6e-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: E1009 08:37:51.387177 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd15ded0_0886_42fd_af2b_24af9dcac27b.slice/crio-conmon-ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd15ded0_0886_42fd_af2b_24af9dcac27b.slice/crio-a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfd15ded0_0886_42fd_af2b_24af9dcac27b.slice/crio-ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b.scope\": RecentStats: unable to find data in memory cache]" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.428299 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.428761 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.189:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.504169 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.689257 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.833537 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs\") pod \"fd15ded0-0886-42fd-af2b-24af9dcac27b\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.833621 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data\") pod \"fd15ded0-0886-42fd-af2b-24af9dcac27b\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.833754 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs\") pod \"fd15ded0-0886-42fd-af2b-24af9dcac27b\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.833902 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle\") pod \"fd15ded0-0886-42fd-af2b-24af9dcac27b\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.833939 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tb5q2\" (UniqueName: \"kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2\") pod \"fd15ded0-0886-42fd-af2b-24af9dcac27b\" (UID: \"fd15ded0-0886-42fd-af2b-24af9dcac27b\") " Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.835682 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs" (OuterVolumeSpecName: "logs") pod "fd15ded0-0886-42fd-af2b-24af9dcac27b" (UID: "fd15ded0-0886-42fd-af2b-24af9dcac27b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.839442 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2" (OuterVolumeSpecName: "kube-api-access-tb5q2") pod "fd15ded0-0886-42fd-af2b-24af9dcac27b" (UID: "fd15ded0-0886-42fd-af2b-24af9dcac27b"). InnerVolumeSpecName "kube-api-access-tb5q2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.869105 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data" (OuterVolumeSpecName: "config-data") pod "fd15ded0-0886-42fd-af2b-24af9dcac27b" (UID: "fd15ded0-0886-42fd-af2b-24af9dcac27b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.871730 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fd15ded0-0886-42fd-af2b-24af9dcac27b" (UID: "fd15ded0-0886-42fd-af2b-24af9dcac27b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.887088 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "fd15ded0-0886-42fd-af2b-24af9dcac27b" (UID: "fd15ded0-0886-42fd-af2b-24af9dcac27b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927165 4872 generic.go:334] "Generic (PLEG): container finished" podID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerID="a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" exitCode=0 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927193 4872 generic.go:334] "Generic (PLEG): container finished" podID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerID="ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" exitCode=143 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927228 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerDied","Data":"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2"} Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927253 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerDied","Data":"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b"} Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927262 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"fd15ded0-0886-42fd-af2b-24af9dcac27b","Type":"ContainerDied","Data":"220b66ffd75e98c7dc8cfdadcc7dbc7564acf82b32c345fe9c3b1d39fbbbffd0"} Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927276 4872 scope.go:117] "RemoveContainer" containerID="a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.927383 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.935272 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" event={"ID":"3b4436f4-a95e-4b70-9e13-02f70d26ee6e","Type":"ContainerDied","Data":"278f3f0426596fa38aaa36894a5b9d7c2b9780f778fe1431c0fb812e166bebec"} Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.935362 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bb4fc677f-glltb" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.935446 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-log" containerID="cri-o://ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1" gracePeriod=30 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.935562 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-api" containerID="cri-o://1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a" gracePeriod=30 Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.937548 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.937578 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tb5q2\" (UniqueName: \"kubernetes.io/projected/fd15ded0-0886-42fd-af2b-24af9dcac27b-kube-api-access-tb5q2\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.937592 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd15ded0-0886-42fd-af2b-24af9dcac27b-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.937605 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.937616 4872 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd15ded0-0886-42fd-af2b-24af9dcac27b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.981818 4872 scope.go:117] "RemoveContainer" containerID="ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" Oct 09 08:37:51 crc kubenswrapper[4872]: I1009 08:37:51.998881 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.011158 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.020720 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.031128 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bb4fc677f-glltb"] Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.041497 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.042265 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96070559-6af7-4827-8dd7-38edfdd2ac47" containerName="nova-manage" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042301 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="96070559-6af7-4827-8dd7-38edfdd2ac47" containerName="nova-manage" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.042310 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-metadata" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042316 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-metadata" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.042334 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="init" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042340 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="init" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.042378 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="dnsmasq-dns" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042384 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="dnsmasq-dns" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.042394 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-log" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042400 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-log" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042773 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" containerName="dnsmasq-dns" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042796 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-log" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042809 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" containerName="nova-metadata-metadata" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.042853 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="96070559-6af7-4827-8dd7-38edfdd2ac47" containerName="nova-manage" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.059825 4872 scope.go:117] "RemoveContainer" containerID="a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.071887 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2\": container with ID starting with a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2 not found: ID does not exist" containerID="a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.071979 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2"} err="failed to get container status \"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2\": rpc error: code = NotFound desc = could not find container \"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2\": container with ID starting with a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2 not found: ID does not exist" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.072014 4872 scope.go:117] "RemoveContainer" containerID="ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" Oct 09 08:37:52 crc kubenswrapper[4872]: E1009 08:37:52.072770 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b\": container with ID starting with ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b not found: ID does not exist" containerID="ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.077132 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b"} err="failed to get container status \"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b\": rpc error: code = NotFound desc = could not find container \"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b\": container with ID starting with ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b not found: ID does not exist" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.077180 4872 scope.go:117] "RemoveContainer" containerID="a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.082330 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2"} err="failed to get container status \"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2\": rpc error: code = NotFound desc = could not find container \"a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2\": container with ID starting with a58b7cb56c7c9f417a1ac0425b7df55b9ad5ec91e290f2ac2431e1319db24ac2 not found: ID does not exist" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.082378 4872 scope.go:117] "RemoveContainer" containerID="ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.082516 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.091548 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.092416 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b"} err="failed to get container status \"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b\": rpc error: code = NotFound desc = could not find container \"ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b\": container with ID starting with ca6268fc7c4d486bae0c03c3501c6bd6fae0097be3ff5a4d66e89b5b6574d85b not found: ID does not exist" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.092454 4872 scope.go:117] "RemoveContainer" containerID="dfe5352f1ca2fd7daae4cceb44cada175049f883c7a926b8c9de8f52508e12c6" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.095522 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.095848 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.131462 4872 scope.go:117] "RemoveContainer" containerID="e20bd1006477608efcb7b91bd3cf777cb95ea68a5f7e43665d742b83925cd44f" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.141467 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc6ff\" (UniqueName: \"kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.141568 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.141729 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.141803 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.141965 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.244354 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc6ff\" (UniqueName: \"kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.244825 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.244946 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.245007 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.245175 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.245584 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.252845 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.254250 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.258016 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.268225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc6ff\" (UniqueName: \"kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff\") pod \"nova-metadata-0\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.347467 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.418331 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.448312 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data\") pod \"3b76157d-7dc5-4661-972e-6673f3bd9990\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.448442 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts\") pod \"3b76157d-7dc5-4661-972e-6673f3bd9990\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.448478 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvl4b\" (UniqueName: \"kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b\") pod \"3b76157d-7dc5-4661-972e-6673f3bd9990\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.448527 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle\") pod \"3b76157d-7dc5-4661-972e-6673f3bd9990\" (UID: \"3b76157d-7dc5-4661-972e-6673f3bd9990\") " Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.453495 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts" (OuterVolumeSpecName: "scripts") pod "3b76157d-7dc5-4661-972e-6673f3bd9990" (UID: "3b76157d-7dc5-4661-972e-6673f3bd9990"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.457337 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b" (OuterVolumeSpecName: "kube-api-access-pvl4b") pod "3b76157d-7dc5-4661-972e-6673f3bd9990" (UID: "3b76157d-7dc5-4661-972e-6673f3bd9990"). InnerVolumeSpecName "kube-api-access-pvl4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.480778 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b4436f4-a95e-4b70-9e13-02f70d26ee6e" path="/var/lib/kubelet/pods/3b4436f4-a95e-4b70-9e13-02f70d26ee6e/volumes" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.481780 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd15ded0-0886-42fd-af2b-24af9dcac27b" path="/var/lib/kubelet/pods/fd15ded0-0886-42fd-af2b-24af9dcac27b/volumes" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.490904 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3b76157d-7dc5-4661-972e-6673f3bd9990" (UID: "3b76157d-7dc5-4661-972e-6673f3bd9990"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.491894 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data" (OuterVolumeSpecName: "config-data") pod "3b76157d-7dc5-4661-972e-6673f3bd9990" (UID: "3b76157d-7dc5-4661-972e-6673f3bd9990"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.552184 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.552549 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.552563 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3b76157d-7dc5-4661-972e-6673f3bd9990-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.552577 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvl4b\" (UniqueName: \"kubernetes.io/projected/3b76157d-7dc5-4661-972e-6673f3bd9990-kube-api-access-pvl4b\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.901246 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:37:52 crc kubenswrapper[4872]: W1009 08:37:52.903951 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10f65ca4_ef8b_40de_88eb_e4e42f78e553.slice/crio-38a262cd0592822345d5737affbae53bb8b937f8bbe242657e1fda8716ef9e87 WatchSource:0}: Error finding container 38a262cd0592822345d5737affbae53bb8b937f8bbe242657e1fda8716ef9e87: Status 404 returned error can't find the container with id 38a262cd0592822345d5737affbae53bb8b937f8bbe242657e1fda8716ef9e87 Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.949155 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.949152 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-hrvkd" event={"ID":"3b76157d-7dc5-4661-972e-6673f3bd9990","Type":"ContainerDied","Data":"7fba23fb5bb0b8f7c4067e76f141c606a0f6c9f458cd00ba9edec12663017511"} Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.949282 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fba23fb5bb0b8f7c4067e76f141c606a0f6c9f458cd00ba9edec12663017511" Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.950809 4872 generic.go:334] "Generic (PLEG): container finished" podID="352f038d-8e90-47cb-9522-89df059086a9" containerID="ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1" exitCode=143 Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.950868 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerDied","Data":"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1"} Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.952627 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerStarted","Data":"38a262cd0592822345d5737affbae53bb8b937f8bbe242657e1fda8716ef9e87"} Oct 09 08:37:52 crc kubenswrapper[4872]: I1009 08:37:52.953858 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" containerName="nova-scheduler-scheduler" containerID="cri-o://7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" gracePeriod=30 Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.012236 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 08:37:53 crc kubenswrapper[4872]: E1009 08:37:53.012892 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b76157d-7dc5-4661-972e-6673f3bd9990" containerName="nova-cell1-conductor-db-sync" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.012906 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b76157d-7dc5-4661-972e-6673f3bd9990" containerName="nova-cell1-conductor-db-sync" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.013063 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b76157d-7dc5-4661-972e-6673f3bd9990" containerName="nova-cell1-conductor-db-sync" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.013652 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.022879 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.024720 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.060481 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.060631 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.060673 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntmks\" (UniqueName: \"kubernetes.io/projected/cf29706b-8cae-42b7-ba76-b52189349378-kube-api-access-ntmks\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.164393 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.164587 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.164623 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntmks\" (UniqueName: \"kubernetes.io/projected/cf29706b-8cae-42b7-ba76-b52189349378-kube-api-access-ntmks\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.170009 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.174946 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf29706b-8cae-42b7-ba76-b52189349378-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.185259 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntmks\" (UniqueName: \"kubernetes.io/projected/cf29706b-8cae-42b7-ba76-b52189349378-kube-api-access-ntmks\") pod \"nova-cell1-conductor-0\" (UID: \"cf29706b-8cae-42b7-ba76-b52189349378\") " pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.327948 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: W1009 08:37:53.775076 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf29706b_8cae_42b7_ba76_b52189349378.slice/crio-c00daa39b3fcb5a0145be8085f126e4b1f43710fb89dd9105f08eabce1fd7681 WatchSource:0}: Error finding container c00daa39b3fcb5a0145be8085f126e4b1f43710fb89dd9105f08eabce1fd7681: Status 404 returned error can't find the container with id c00daa39b3fcb5a0145be8085f126e4b1f43710fb89dd9105f08eabce1fd7681 Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.778722 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.969014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cf29706b-8cae-42b7-ba76-b52189349378","Type":"ContainerStarted","Data":"90d3b9d5d33b7dbed95ed5eb6e4a0f311c878145aa55206727f97d489dae61d4"} Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.969063 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"cf29706b-8cae-42b7-ba76-b52189349378","Type":"ContainerStarted","Data":"c00daa39b3fcb5a0145be8085f126e4b1f43710fb89dd9105f08eabce1fd7681"} Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.969133 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.971442 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerStarted","Data":"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4"} Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.971474 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerStarted","Data":"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3"} Oct 09 08:37:53 crc kubenswrapper[4872]: I1009 08:37:53.989374 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=1.9893528219999999 podStartE2EDuration="1.989352822s" podCreationTimestamp="2025-10-09 08:37:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:53.984383069 +0000 UTC m=+1232.174911705" watchObservedRunningTime="2025-10-09 08:37:53.989352822 +0000 UTC m=+1232.179881458" Oct 09 08:37:54 crc kubenswrapper[4872]: I1009 08:37:54.010768 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.01074735 podStartE2EDuration="3.01074735s" podCreationTimestamp="2025-10-09 08:37:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:54.001108892 +0000 UTC m=+1232.191637528" watchObservedRunningTime="2025-10-09 08:37:54.01074735 +0000 UTC m=+1232.201275976" Oct 09 08:37:55 crc kubenswrapper[4872]: E1009 08:37:55.304206 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 08:37:55 crc kubenswrapper[4872]: E1009 08:37:55.306362 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 08:37:55 crc kubenswrapper[4872]: E1009 08:37:55.307994 4872 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Oct 09 08:37:55 crc kubenswrapper[4872]: E1009 08:37:55.308066 4872 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" containerName="nova-scheduler-scheduler" Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.891725 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.998752 4872 generic.go:334] "Generic (PLEG): container finished" podID="7860ad5a-5664-4bc7-8847-146479f7d036" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" exitCode=0 Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.998795 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ad5a-5664-4bc7-8847-146479f7d036","Type":"ContainerDied","Data":"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0"} Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.998827 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7860ad5a-5664-4bc7-8847-146479f7d036","Type":"ContainerDied","Data":"9ee1ca6a11e32098e8b2a8004ec3318f32f6406df03e52282f1c7d42257f6761"} Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.998830 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:56 crc kubenswrapper[4872]: I1009 08:37:56.998843 4872 scope.go:117] "RemoveContainer" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.019913 4872 scope.go:117] "RemoveContainer" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" Oct 09 08:37:57 crc kubenswrapper[4872]: E1009 08:37:57.020694 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0\": container with ID starting with 7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0 not found: ID does not exist" containerID="7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.020747 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0"} err="failed to get container status \"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0\": rpc error: code = NotFound desc = could not find container \"7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0\": container with ID starting with 7cd6eedd63477c6cda2120a62a688015348bc3117158a2b41aa6f50227e7eec0 not found: ID does not exist" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.048701 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle\") pod \"7860ad5a-5664-4bc7-8847-146479f7d036\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.048768 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq5bf\" (UniqueName: \"kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf\") pod \"7860ad5a-5664-4bc7-8847-146479f7d036\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.048810 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data\") pod \"7860ad5a-5664-4bc7-8847-146479f7d036\" (UID: \"7860ad5a-5664-4bc7-8847-146479f7d036\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.063901 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf" (OuterVolumeSpecName: "kube-api-access-tq5bf") pod "7860ad5a-5664-4bc7-8847-146479f7d036" (UID: "7860ad5a-5664-4bc7-8847-146479f7d036"). InnerVolumeSpecName "kube-api-access-tq5bf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.074973 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data" (OuterVolumeSpecName: "config-data") pod "7860ad5a-5664-4bc7-8847-146479f7d036" (UID: "7860ad5a-5664-4bc7-8847-146479f7d036"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.078149 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7860ad5a-5664-4bc7-8847-146479f7d036" (UID: "7860ad5a-5664-4bc7-8847-146479f7d036"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.150806 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.151044 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq5bf\" (UniqueName: \"kubernetes.io/projected/7860ad5a-5664-4bc7-8847-146479f7d036-kube-api-access-tq5bf\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.151055 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7860ad5a-5664-4bc7-8847-146479f7d036-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.333189 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.341706 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.360862 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:57 crc kubenswrapper[4872]: E1009 08:37:57.361208 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" containerName="nova-scheduler-scheduler" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.361228 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" containerName="nova-scheduler-scheduler" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.361494 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" containerName="nova-scheduler-scheduler" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.362087 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.370712 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.375015 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.418811 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.418970 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.458157 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.458209 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.458369 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xqlb\" (UniqueName: \"kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.560588 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xqlb\" (UniqueName: \"kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.560827 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.560851 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.564916 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.565892 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.595366 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xqlb\" (UniqueName: \"kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb\") pod \"nova-scheduler-0\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.728615 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.824618 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.968028 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle\") pod \"352f038d-8e90-47cb-9522-89df059086a9\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.968539 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs\") pod \"352f038d-8e90-47cb-9522-89df059086a9\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.968666 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxmqb\" (UniqueName: \"kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb\") pod \"352f038d-8e90-47cb-9522-89df059086a9\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.968700 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data\") pod \"352f038d-8e90-47cb-9522-89df059086a9\" (UID: \"352f038d-8e90-47cb-9522-89df059086a9\") " Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.970323 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs" (OuterVolumeSpecName: "logs") pod "352f038d-8e90-47cb-9522-89df059086a9" (UID: "352f038d-8e90-47cb-9522-89df059086a9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.971115 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/352f038d-8e90-47cb-9522-89df059086a9-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:57 crc kubenswrapper[4872]: I1009 08:37:57.976512 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb" (OuterVolumeSpecName: "kube-api-access-wxmqb") pod "352f038d-8e90-47cb-9522-89df059086a9" (UID: "352f038d-8e90-47cb-9522-89df059086a9"). InnerVolumeSpecName "kube-api-access-wxmqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023106 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data" (OuterVolumeSpecName: "config-data") pod "352f038d-8e90-47cb-9522-89df059086a9" (UID: "352f038d-8e90-47cb-9522-89df059086a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023714 4872 generic.go:334] "Generic (PLEG): container finished" podID="352f038d-8e90-47cb-9522-89df059086a9" containerID="1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a" exitCode=0 Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023803 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerDied","Data":"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a"} Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023834 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"352f038d-8e90-47cb-9522-89df059086a9","Type":"ContainerDied","Data":"8ec79e351987ba59012bb49ae931df6b75687db7e22a43b8637184933a9f232e"} Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023851 4872 scope.go:117] "RemoveContainer" containerID="1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023893 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.023717 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "352f038d-8e90-47cb-9522-89df059086a9" (UID: "352f038d-8e90-47cb-9522-89df059086a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.056301 4872 scope.go:117] "RemoveContainer" containerID="ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.059255 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.078855 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxmqb\" (UniqueName: \"kubernetes.io/projected/352f038d-8e90-47cb-9522-89df059086a9-kube-api-access-wxmqb\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.078895 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.078906 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/352f038d-8e90-47cb-9522-89df059086a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.088369 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.099615 4872 scope.go:117] "RemoveContainer" containerID="1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a" Oct 09 08:37:58 crc kubenswrapper[4872]: E1009 08:37:58.104028 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a\": container with ID starting with 1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a not found: ID does not exist" containerID="1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.104082 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a"} err="failed to get container status \"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a\": rpc error: code = NotFound desc = could not find container \"1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a\": container with ID starting with 1997e49471f9ed6dbeae37825fc72a453bab318544dae47c2f878c6ad139bc2a not found: ID does not exist" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.104112 4872 scope.go:117] "RemoveContainer" containerID="ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1" Oct 09 08:37:58 crc kubenswrapper[4872]: E1009 08:37:58.104770 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1\": container with ID starting with ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1 not found: ID does not exist" containerID="ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.104793 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1"} err="failed to get container status \"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1\": rpc error: code = NotFound desc = could not find container \"ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1\": container with ID starting with ee3a2b3084b2bbdced6e981704959e3b46b517e01d624a223d29ac95b12712a1 not found: ID does not exist" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.109244 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:58 crc kubenswrapper[4872]: E1009 08:37:58.109741 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-log" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.109759 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-log" Oct 09 08:37:58 crc kubenswrapper[4872]: E1009 08:37:58.109775 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-api" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.109783 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-api" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.109989 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-api" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.110030 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="352f038d-8e90-47cb-9522-89df059086a9" containerName="nova-api-log" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.111031 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.117717 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.117883 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.181659 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.181726 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxxg4\" (UniqueName: \"kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.183332 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.183784 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.185451 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:37:58 crc kubenswrapper[4872]: W1009 08:37:58.188594 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod15a338c3_8256_4891_a0b8_ffab1db121a5.slice/crio-a95d26dc58438386fa7fb27fddfa97bf5a1237eeee353abc39839a4685946c71 WatchSource:0}: Error finding container a95d26dc58438386fa7fb27fddfa97bf5a1237eeee353abc39839a4685946c71: Status 404 returned error can't find the container with id a95d26dc58438386fa7fb27fddfa97bf5a1237eeee353abc39839a4685946c71 Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.285062 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.285105 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.285129 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxxg4\" (UniqueName: \"kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.285205 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.285663 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.289371 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.290256 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.301544 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxxg4\" (UniqueName: \"kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4\") pod \"nova-api-0\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.364453 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.430622 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.473214 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="352f038d-8e90-47cb-9522-89df059086a9" path="/var/lib/kubelet/pods/352f038d-8e90-47cb-9522-89df059086a9/volumes" Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.473882 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7860ad5a-5664-4bc7-8847-146479f7d036" path="/var/lib/kubelet/pods/7860ad5a-5664-4bc7-8847-146479f7d036/volumes" Oct 09 08:37:58 crc kubenswrapper[4872]: W1009 08:37:58.896728 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f680e1a_99ae_41b7_b6e0_8d9e22fcee6f.slice/crio-64c83675412d7df7c984c57d0b4320f711b8414a9106ece07c72c11a77d734a6 WatchSource:0}: Error finding container 64c83675412d7df7c984c57d0b4320f711b8414a9106ece07c72c11a77d734a6: Status 404 returned error can't find the container with id 64c83675412d7df7c984c57d0b4320f711b8414a9106ece07c72c11a77d734a6 Oct 09 08:37:58 crc kubenswrapper[4872]: I1009 08:37:58.897196 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:37:59 crc kubenswrapper[4872]: I1009 08:37:59.043242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerStarted","Data":"64c83675412d7df7c984c57d0b4320f711b8414a9106ece07c72c11a77d734a6"} Oct 09 08:37:59 crc kubenswrapper[4872]: I1009 08:37:59.047428 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a338c3-8256-4891-a0b8-ffab1db121a5","Type":"ContainerStarted","Data":"25523c74e5ffa42f87f4263e193bfdd893ab2dcd520ea6b60443c8b2ecc6fa65"} Oct 09 08:37:59 crc kubenswrapper[4872]: I1009 08:37:59.047823 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a338c3-8256-4891-a0b8-ffab1db121a5","Type":"ContainerStarted","Data":"a95d26dc58438386fa7fb27fddfa97bf5a1237eeee353abc39839a4685946c71"} Oct 09 08:37:59 crc kubenswrapper[4872]: I1009 08:37:59.068861 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.06884313 podStartE2EDuration="2.06884313s" podCreationTimestamp="2025-10-09 08:37:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:37:59.0632843 +0000 UTC m=+1237.253812926" watchObservedRunningTime="2025-10-09 08:37:59.06884313 +0000 UTC m=+1237.259371996" Oct 09 08:38:00 crc kubenswrapper[4872]: I1009 08:38:00.062600 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerStarted","Data":"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296"} Oct 09 08:38:00 crc kubenswrapper[4872]: I1009 08:38:00.062689 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerStarted","Data":"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae"} Oct 09 08:38:02 crc kubenswrapper[4872]: I1009 08:38:02.418681 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 08:38:02 crc kubenswrapper[4872]: I1009 08:38:02.419116 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 08:38:02 crc kubenswrapper[4872]: I1009 08:38:02.729339 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 08:38:03 crc kubenswrapper[4872]: I1009 08:38:03.431817 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:03 crc kubenswrapper[4872]: I1009 08:38:03.431862 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.193:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:05 crc kubenswrapper[4872]: I1009 08:38:05.955309 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 08:38:06 crc kubenswrapper[4872]: I1009 08:38:06.000873 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=8.000801798 podStartE2EDuration="8.000801798s" podCreationTimestamp="2025-10-09 08:37:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:00.091334463 +0000 UTC m=+1238.281863129" watchObservedRunningTime="2025-10-09 08:38:06.000801798 +0000 UTC m=+1244.191330434" Oct 09 08:38:07 crc kubenswrapper[4872]: I1009 08:38:07.729369 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 08:38:07 crc kubenswrapper[4872]: I1009 08:38:07.758910 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 08:38:08 crc kubenswrapper[4872]: I1009 08:38:08.176561 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 08:38:08 crc kubenswrapper[4872]: I1009 08:38:08.431463 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:08 crc kubenswrapper[4872]: I1009 08:38:08.432758 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.296952 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.297199 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="1effc4cb-04ae-4ee7-8de4-76502c225942" containerName="kube-state-metrics" containerID="cri-o://862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a" gracePeriod=30 Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.517544 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.517561 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.196:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.764130 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.938544 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvb6h\" (UniqueName: \"kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h\") pod \"1effc4cb-04ae-4ee7-8de4-76502c225942\" (UID: \"1effc4cb-04ae-4ee7-8de4-76502c225942\") " Oct 09 08:38:09 crc kubenswrapper[4872]: I1009 08:38:09.949368 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h" (OuterVolumeSpecName: "kube-api-access-pvb6h") pod "1effc4cb-04ae-4ee7-8de4-76502c225942" (UID: "1effc4cb-04ae-4ee7-8de4-76502c225942"). InnerVolumeSpecName "kube-api-access-pvb6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.040322 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvb6h\" (UniqueName: \"kubernetes.io/projected/1effc4cb-04ae-4ee7-8de4-76502c225942-kube-api-access-pvb6h\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.171182 4872 generic.go:334] "Generic (PLEG): container finished" podID="1effc4cb-04ae-4ee7-8de4-76502c225942" containerID="862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a" exitCode=2 Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.171246 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1effc4cb-04ae-4ee7-8de4-76502c225942","Type":"ContainerDied","Data":"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a"} Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.171279 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"1effc4cb-04ae-4ee7-8de4-76502c225942","Type":"ContainerDied","Data":"adaac7315338e7885def7de1a278a6691cc5e03eeb39345b95f09ebb4bf9698c"} Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.171300 4872 scope.go:117] "RemoveContainer" containerID="862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.171423 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.204845 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.205104 4872 scope.go:117] "RemoveContainer" containerID="862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a" Oct 09 08:38:10 crc kubenswrapper[4872]: E1009 08:38:10.205551 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a\": container with ID starting with 862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a not found: ID does not exist" containerID="862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.205594 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a"} err="failed to get container status \"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a\": rpc error: code = NotFound desc = could not find container \"862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a\": container with ID starting with 862c09d07d001038ad3e61fa66fe6e8cf302af34aefc2d2aeb7d0deff0a7bc3a not found: ID does not exist" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.213743 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.235935 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:10 crc kubenswrapper[4872]: E1009 08:38:10.236433 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1effc4cb-04ae-4ee7-8de4-76502c225942" containerName="kube-state-metrics" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.236458 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="1effc4cb-04ae-4ee7-8de4-76502c225942" containerName="kube-state-metrics" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.236711 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="1effc4cb-04ae-4ee7-8de4-76502c225942" containerName="kube-state-metrics" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.238894 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.241028 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.241297 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.245233 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.344624 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.344717 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.344811 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.344881 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2h2s8\" (UniqueName: \"kubernetes.io/projected/4d540b3a-5230-484c-b807-ce072ce235d9-kube-api-access-2h2s8\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.446772 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.446854 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.446920 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.446975 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2h2s8\" (UniqueName: \"kubernetes.io/projected/4d540b3a-5230-484c-b807-ce072ce235d9-kube-api-access-2h2s8\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.450670 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.453462 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.456272 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/4d540b3a-5230-484c-b807-ce072ce235d9-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.472868 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2h2s8\" (UniqueName: \"kubernetes.io/projected/4d540b3a-5230-484c-b807-ce072ce235d9-kube-api-access-2h2s8\") pod \"kube-state-metrics-0\" (UID: \"4d540b3a-5230-484c-b807-ce072ce235d9\") " pod="openstack/kube-state-metrics-0" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.480792 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1effc4cb-04ae-4ee7-8de4-76502c225942" path="/var/lib/kubelet/pods/1effc4cb-04ae-4ee7-8de4-76502c225942/volumes" Oct 09 08:38:10 crc kubenswrapper[4872]: I1009 08:38:10.565834 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.078010 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 09 08:38:11 crc kubenswrapper[4872]: W1009 08:38:11.090297 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d540b3a_5230_484c_b807_ce072ce235d9.slice/crio-79bab145b81be99c4d929821d9514386b115d3bed9792e8030750fbd9a26f613 WatchSource:0}: Error finding container 79bab145b81be99c4d929821d9514386b115d3bed9792e8030750fbd9a26f613: Status 404 returned error can't find the container with id 79bab145b81be99c4d929821d9514386b115d3bed9792e8030750fbd9a26f613 Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.183221 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d540b3a-5230-484c-b807-ce072ce235d9","Type":"ContainerStarted","Data":"79bab145b81be99c4d929821d9514386b115d3bed9792e8030750fbd9a26f613"} Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.210356 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.210621 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-central-agent" containerID="cri-o://3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41" gracePeriod=30 Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.210790 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="proxy-httpd" containerID="cri-o://371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53" gracePeriod=30 Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.210856 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-notification-agent" containerID="cri-o://07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0" gracePeriod=30 Oct 09 08:38:11 crc kubenswrapper[4872]: I1009 08:38:11.211011 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="sg-core" containerID="cri-o://1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799" gracePeriod=30 Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.193043 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"4d540b3a-5230-484c-b807-ce072ce235d9","Type":"ContainerStarted","Data":"d3e17cdf07d588e828d24f64dcc74d57d3681e865b9b638631a44ac0e0442e4f"} Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.193350 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196189 4872 generic.go:334] "Generic (PLEG): container finished" podID="14114366-3880-4288-9d9c-ffc0d7610046" containerID="371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53" exitCode=0 Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196226 4872 generic.go:334] "Generic (PLEG): container finished" podID="14114366-3880-4288-9d9c-ffc0d7610046" containerID="1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799" exitCode=2 Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196237 4872 generic.go:334] "Generic (PLEG): container finished" podID="14114366-3880-4288-9d9c-ffc0d7610046" containerID="3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41" exitCode=0 Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196258 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerDied","Data":"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53"} Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196296 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerDied","Data":"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799"} Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.196307 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerDied","Data":"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41"} Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.222502 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.8672228199999998 podStartE2EDuration="2.222481054s" podCreationTimestamp="2025-10-09 08:38:10 +0000 UTC" firstStartedPulling="2025-10-09 08:38:11.093830996 +0000 UTC m=+1249.284359622" lastFinishedPulling="2025-10-09 08:38:11.44908922 +0000 UTC m=+1249.639617856" observedRunningTime="2025-10-09 08:38:12.216730708 +0000 UTC m=+1250.407259354" watchObservedRunningTime="2025-10-09 08:38:12.222481054 +0000 UTC m=+1250.413009690" Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.425667 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.426005 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.430119 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 08:38:12 crc kubenswrapper[4872]: I1009 08:38:12.435340 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 08:38:14 crc kubenswrapper[4872]: I1009 08:38:14.943717 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033510 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033700 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033745 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdb2n\" (UniqueName: \"kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033783 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033862 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033886 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.033948 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml\") pod \"14114366-3880-4288-9d9c-ffc0d7610046\" (UID: \"14114366-3880-4288-9d9c-ffc0d7610046\") " Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.034757 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.034945 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.035422 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.041188 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n" (OuterVolumeSpecName: "kube-api-access-gdb2n") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "kube-api-access-gdb2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.056865 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts" (OuterVolumeSpecName: "scripts") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.064802 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.127119 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.136328 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdb2n\" (UniqueName: \"kubernetes.io/projected/14114366-3880-4288-9d9c-ffc0d7610046-kube-api-access-gdb2n\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.136361 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.136371 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/14114366-3880-4288-9d9c-ffc0d7610046-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.136383 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.136393 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.151205 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data" (OuterVolumeSpecName: "config-data") pod "14114366-3880-4288-9d9c-ffc0d7610046" (UID: "14114366-3880-4288-9d9c-ffc0d7610046"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.227812 4872 generic.go:334] "Generic (PLEG): container finished" podID="14114366-3880-4288-9d9c-ffc0d7610046" containerID="07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0" exitCode=0 Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.227883 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerDied","Data":"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0"} Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.227934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"14114366-3880-4288-9d9c-ffc0d7610046","Type":"ContainerDied","Data":"db2e9f4e76c5cf40fc4d78c753687b8d2f3adfb3cdf74962c6f1aa84693aa196"} Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.227951 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.227963 4872 scope.go:117] "RemoveContainer" containerID="371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.237795 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/14114366-3880-4288-9d9c-ffc0d7610046-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.269906 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.276301 4872 scope.go:117] "RemoveContainer" containerID="1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.294090 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.323917 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.324720 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-notification-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.324744 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-notification-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.324779 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="proxy-httpd" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.324789 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="proxy-httpd" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.324820 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="sg-core" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.324829 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="sg-core" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.324874 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-central-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.324884 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-central-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.330499 4872 scope.go:117] "RemoveContainer" containerID="07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.330589 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="proxy-httpd" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.330622 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="sg-core" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.330658 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-central-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.330717 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14114366-3880-4288-9d9c-ffc0d7610046" containerName="ceilometer-notification-agent" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.339106 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.341047 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.343919 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.344151 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.344258 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.362704 4872 scope.go:117] "RemoveContainer" containerID="3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.383991 4872 scope.go:117] "RemoveContainer" containerID="371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.384897 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53\": container with ID starting with 371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53 not found: ID does not exist" containerID="371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.385042 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53"} err="failed to get container status \"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53\": rpc error: code = NotFound desc = could not find container \"371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53\": container with ID starting with 371c42d810a123c37250f0cd85579470fcba07f51d37662ea1e1e76b56a67e53 not found: ID does not exist" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.385141 4872 scope.go:117] "RemoveContainer" containerID="1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.385676 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799\": container with ID starting with 1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799 not found: ID does not exist" containerID="1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.385788 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799"} err="failed to get container status \"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799\": rpc error: code = NotFound desc = could not find container \"1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799\": container with ID starting with 1d4725f522cbcff2a55d90d1907367fea27027856a64e1f9f6e6cab8d2aa9799 not found: ID does not exist" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.385926 4872 scope.go:117] "RemoveContainer" containerID="07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.386259 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0\": container with ID starting with 07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0 not found: ID does not exist" containerID="07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.386345 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0"} err="failed to get container status \"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0\": rpc error: code = NotFound desc = could not find container \"07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0\": container with ID starting with 07411a47fa955ec408b2c4369361c66180fec2451573e99b426e5fd84f18d7e0 not found: ID does not exist" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.386417 4872 scope.go:117] "RemoveContainer" containerID="3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41" Oct 09 08:38:15 crc kubenswrapper[4872]: E1009 08:38:15.386674 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41\": container with ID starting with 3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41 not found: ID does not exist" containerID="3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.386758 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41"} err="failed to get container status \"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41\": rpc error: code = NotFound desc = could not find container \"3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41\": container with ID starting with 3128334abcdb45e9f8a87addf6c978d74153af7e242a44a1fd04a481443c4a41 not found: ID does not exist" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442260 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442375 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442429 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442456 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442552 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442576 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442605 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.442640 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5wws\" (UniqueName: \"kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544684 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544748 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544827 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5wws\" (UniqueName: \"kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544881 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544944 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.544982 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.545007 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.545587 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.545828 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.549227 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.549227 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.550866 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.557957 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.558201 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.560805 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5wws\" (UniqueName: \"kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws\") pod \"ceilometer-0\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " pod="openstack/ceilometer-0" Oct 09 08:38:15 crc kubenswrapper[4872]: I1009 08:38:15.663850 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.123649 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:16 crc kubenswrapper[4872]: W1009 08:38:16.149947 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e WatchSource:0}: Error finding container 55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e: Status 404 returned error can't find the container with id 55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.240800 4872 generic.go:334] "Generic (PLEG): container finished" podID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" containerID="e5456a0bc88528481d0e92e61aeb1fa5c43395c2ec951ebab383d1b09cac6bf3" exitCode=137 Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.240910 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2336b0ba-f01d-4013-a125-dbfc3a5c384d","Type":"ContainerDied","Data":"e5456a0bc88528481d0e92e61aeb1fa5c43395c2ec951ebab383d1b09cac6bf3"} Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.240973 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2336b0ba-f01d-4013-a125-dbfc3a5c384d","Type":"ContainerDied","Data":"7945151881e59ff81aa5dc335fb5e86ef72147f1528b7b8a0445382688e130b5"} Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.240984 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7945151881e59ff81aa5dc335fb5e86ef72147f1528b7b8a0445382688e130b5" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.244128 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerStarted","Data":"55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e"} Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.244355 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.366465 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wlb7\" (UniqueName: \"kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7\") pod \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.366573 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle\") pod \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.366901 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data\") pod \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\" (UID: \"2336b0ba-f01d-4013-a125-dbfc3a5c384d\") " Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.372007 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7" (OuterVolumeSpecName: "kube-api-access-9wlb7") pod "2336b0ba-f01d-4013-a125-dbfc3a5c384d" (UID: "2336b0ba-f01d-4013-a125-dbfc3a5c384d"). InnerVolumeSpecName "kube-api-access-9wlb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.396427 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data" (OuterVolumeSpecName: "config-data") pod "2336b0ba-f01d-4013-a125-dbfc3a5c384d" (UID: "2336b0ba-f01d-4013-a125-dbfc3a5c384d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.415250 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2336b0ba-f01d-4013-a125-dbfc3a5c384d" (UID: "2336b0ba-f01d-4013-a125-dbfc3a5c384d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.469074 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.469116 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wlb7\" (UniqueName: \"kubernetes.io/projected/2336b0ba-f01d-4013-a125-dbfc3a5c384d-kube-api-access-9wlb7\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.469152 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2336b0ba-f01d-4013-a125-dbfc3a5c384d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:16 crc kubenswrapper[4872]: I1009 08:38:16.472266 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14114366-3880-4288-9d9c-ffc0d7610046" path="/var/lib/kubelet/pods/14114366-3880-4288-9d9c-ffc0d7610046/volumes" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.258222 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerStarted","Data":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.258236 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.277646 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.286806 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.306336 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:38:17 crc kubenswrapper[4872]: E1009 08:38:17.306847 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.306873 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.307121 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" containerName="nova-cell1-novncproxy-novncproxy" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.308177 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.311091 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.311168 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.314355 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.318707 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.386507 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.386583 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.386894 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.386995 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.387047 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ht4ch\" (UniqueName: \"kubernetes.io/projected/15c4d286-8c8e-40a2-b034-77f11abf9a08-kube-api-access-ht4ch\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.488408 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.488472 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.488523 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ht4ch\" (UniqueName: \"kubernetes.io/projected/15c4d286-8c8e-40a2-b034-77f11abf9a08-kube-api-access-ht4ch\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.488576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.488607 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.494400 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.494432 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.496189 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.498045 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/15c4d286-8c8e-40a2-b034-77f11abf9a08-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.507829 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ht4ch\" (UniqueName: \"kubernetes.io/projected/15c4d286-8c8e-40a2-b034-77f11abf9a08-kube-api-access-ht4ch\") pod \"nova-cell1-novncproxy-0\" (UID: \"15c4d286-8c8e-40a2-b034-77f11abf9a08\") " pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:17 crc kubenswrapper[4872]: I1009 08:38:17.634619 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.108080 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.269084 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerStarted","Data":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.270544 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15c4d286-8c8e-40a2-b034-77f11abf9a08","Type":"ContainerStarted","Data":"ef744c90df500769aab5c494be807513be7ee38f3bbbbce9cfc94316785e6204"} Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.435467 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.436041 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.436665 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.438205 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 08:38:18 crc kubenswrapper[4872]: I1009 08:38:18.476285 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2336b0ba-f01d-4013-a125-dbfc3a5c384d" path="/var/lib/kubelet/pods/2336b0ba-f01d-4013-a125-dbfc3a5c384d/volumes" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.281032 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"15c4d286-8c8e-40a2-b034-77f11abf9a08","Type":"ContainerStarted","Data":"628c4d18b34a5c3a1157c0c3a627d3ad30f349857ef3d35288352a4b8e5c61a9"} Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.283473 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerStarted","Data":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.283858 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.296242 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.303924 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.303902357 podStartE2EDuration="2.303902357s" podCreationTimestamp="2025-10-09 08:38:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:19.295816954 +0000 UTC m=+1257.486345580" watchObservedRunningTime="2025-10-09 08:38:19.303902357 +0000 UTC m=+1257.494431003" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.465042 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.467781 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.482761 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.540730 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.540828 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shl2j\" (UniqueName: \"kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.540933 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.540957 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.540982 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.541019 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.642745 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.642793 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.642823 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.642861 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.643202 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.645097 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shl2j\" (UniqueName: \"kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.645342 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.645505 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.646031 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.646451 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.646530 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.664513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shl2j\" (UniqueName: \"kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j\") pod \"dnsmasq-dns-5c7b6c5df9-d4k6h\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:19 crc kubenswrapper[4872]: I1009 08:38:19.798834 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:20 crc kubenswrapper[4872]: I1009 08:38:20.296478 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerStarted","Data":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} Oct 09 08:38:20 crc kubenswrapper[4872]: I1009 08:38:20.332490 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:38:20 crc kubenswrapper[4872]: I1009 08:38:20.332857 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.859941283 podStartE2EDuration="5.332844778s" podCreationTimestamp="2025-10-09 08:38:15 +0000 UTC" firstStartedPulling="2025-10-09 08:38:16.151908606 +0000 UTC m=+1254.342437232" lastFinishedPulling="2025-10-09 08:38:19.624812101 +0000 UTC m=+1257.815340727" observedRunningTime="2025-10-09 08:38:20.329083199 +0000 UTC m=+1258.519611825" watchObservedRunningTime="2025-10-09 08:38:20.332844778 +0000 UTC m=+1258.523373424" Oct 09 08:38:20 crc kubenswrapper[4872]: I1009 08:38:20.574913 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 09 08:38:21 crc kubenswrapper[4872]: I1009 08:38:21.310815 4872 generic.go:334] "Generic (PLEG): container finished" podID="a65d363b-6367-4f29-9d54-9270f6122112" containerID="8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d" exitCode=0 Oct 09 08:38:21 crc kubenswrapper[4872]: I1009 08:38:21.310858 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" event={"ID":"a65d363b-6367-4f29-9d54-9270f6122112","Type":"ContainerDied","Data":"8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d"} Oct 09 08:38:21 crc kubenswrapper[4872]: I1009 08:38:21.311420 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" event={"ID":"a65d363b-6367-4f29-9d54-9270f6122112","Type":"ContainerStarted","Data":"22bf5d6b2c61e3ec660057366f6465f2b8916352768bc69856ff50c2520ce637"} Oct 09 08:38:21 crc kubenswrapper[4872]: I1009 08:38:21.312061 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:38:21 crc kubenswrapper[4872]: I1009 08:38:21.961015 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.321394 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" event={"ID":"a65d363b-6367-4f29-9d54-9270f6122112","Type":"ContainerStarted","Data":"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e"} Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.321474 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-log" containerID="cri-o://8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae" gracePeriod=30 Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.321756 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-api" containerID="cri-o://65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296" gracePeriod=30 Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.348605 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" podStartSLOduration=3.348583781 podStartE2EDuration="3.348583781s" podCreationTimestamp="2025-10-09 08:38:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:22.343394711 +0000 UTC m=+1260.533923327" watchObservedRunningTime="2025-10-09 08:38:22.348583781 +0000 UTC m=+1260.539112407" Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.397832 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:22 crc kubenswrapper[4872]: I1009 08:38:22.635729 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.335069 4872 generic.go:334] "Generic (PLEG): container finished" podID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerID="8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae" exitCode=143 Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.335874 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerDied","Data":"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae"} Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.335905 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.336089 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-central-agent" containerID="cri-o://bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" gracePeriod=30 Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.336606 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="proxy-httpd" containerID="cri-o://0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" gracePeriod=30 Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.336691 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="sg-core" containerID="cri-o://d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" gracePeriod=30 Oct 09 08:38:23 crc kubenswrapper[4872]: I1009 08:38:23.336738 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-notification-agent" containerID="cri-o://f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" gracePeriod=30 Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.189085 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.272589 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.272738 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.272763 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.272856 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.272915 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.273384 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.273556 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.279198 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts" (OuterVolumeSpecName: "scripts") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.306760 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355032 4872 generic.go:334] "Generic (PLEG): container finished" podID="53077894-a0c8-488c-9e7b-794ec805163f" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" exitCode=0 Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355067 4872 generic.go:334] "Generic (PLEG): container finished" podID="53077894-a0c8-488c-9e7b-794ec805163f" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" exitCode=2 Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355074 4872 generic.go:334] "Generic (PLEG): container finished" podID="53077894-a0c8-488c-9e7b-794ec805163f" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" exitCode=0 Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355081 4872 generic.go:334] "Generic (PLEG): container finished" podID="53077894-a0c8-488c-9e7b-794ec805163f" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" exitCode=0 Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355153 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355198 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerDied","Data":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355230 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerDied","Data":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355242 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerDied","Data":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355252 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerDied","Data":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355261 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"53077894-a0c8-488c-9e7b-794ec805163f","Type":"ContainerDied","Data":"55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e"} Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.355277 4872 scope.go:117] "RemoveContainer" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.388106 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5wws\" (UniqueName: \"kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.388163 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.388247 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle\") pod \"53077894-a0c8-488c-9e7b-794ec805163f\" (UID: \"53077894-a0c8-488c-9e7b-794ec805163f\") " Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.390110 4872 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.390247 4872 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.390278 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.390291 4872 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/53077894-a0c8-488c-9e7b-794ec805163f-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.392074 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws" (OuterVolumeSpecName: "kube-api-access-c5wws") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "kube-api-access-c5wws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.409455 4872 scope.go:117] "RemoveContainer" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.422594 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data" (OuterVolumeSpecName: "config-data") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.433892 4872 scope.go:117] "RemoveContainer" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.451831 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.457503 4872 scope.go:117] "RemoveContainer" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.463753 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53077894-a0c8-488c-9e7b-794ec805163f" (UID: "53077894-a0c8-488c-9e7b-794ec805163f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.495107 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5wws\" (UniqueName: \"kubernetes.io/projected/53077894-a0c8-488c-9e7b-794ec805163f-kube-api-access-c5wws\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.495251 4872 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.495337 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.495399 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53077894-a0c8-488c-9e7b-794ec805163f-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.564175 4872 scope.go:117] "RemoveContainer" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.564703 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": container with ID starting with 0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00 not found: ID does not exist" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.564743 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} err="failed to get container status \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": rpc error: code = NotFound desc = could not find container \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": container with ID starting with 0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.564770 4872 scope.go:117] "RemoveContainer" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.565125 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": container with ID starting with d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246 not found: ID does not exist" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565153 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} err="failed to get container status \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": rpc error: code = NotFound desc = could not find container \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": container with ID starting with d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565170 4872 scope.go:117] "RemoveContainer" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.565426 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": container with ID starting with f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1 not found: ID does not exist" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565454 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} err="failed to get container status \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": rpc error: code = NotFound desc = could not find container \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": container with ID starting with f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565472 4872 scope.go:117] "RemoveContainer" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.565686 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": container with ID starting with bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd not found: ID does not exist" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565706 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} err="failed to get container status \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": rpc error: code = NotFound desc = could not find container \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": container with ID starting with bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565721 4872 scope.go:117] "RemoveContainer" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565959 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} err="failed to get container status \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": rpc error: code = NotFound desc = could not find container \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": container with ID starting with 0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.565989 4872 scope.go:117] "RemoveContainer" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.566590 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} err="failed to get container status \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": rpc error: code = NotFound desc = could not find container \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": container with ID starting with d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.566704 4872 scope.go:117] "RemoveContainer" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567094 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} err="failed to get container status \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": rpc error: code = NotFound desc = could not find container \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": container with ID starting with f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567125 4872 scope.go:117] "RemoveContainer" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567383 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} err="failed to get container status \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": rpc error: code = NotFound desc = could not find container \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": container with ID starting with bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567411 4872 scope.go:117] "RemoveContainer" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567745 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} err="failed to get container status \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": rpc error: code = NotFound desc = could not find container \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": container with ID starting with 0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.567779 4872 scope.go:117] "RemoveContainer" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568024 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} err="failed to get container status \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": rpc error: code = NotFound desc = could not find container \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": container with ID starting with d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568047 4872 scope.go:117] "RemoveContainer" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568268 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} err="failed to get container status \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": rpc error: code = NotFound desc = could not find container \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": container with ID starting with f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568284 4872 scope.go:117] "RemoveContainer" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568593 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} err="failed to get container status \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": rpc error: code = NotFound desc = could not find container \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": container with ID starting with bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.568679 4872 scope.go:117] "RemoveContainer" containerID="0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569005 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00"} err="failed to get container status \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": rpc error: code = NotFound desc = could not find container \"0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00\": container with ID starting with 0211c46d70f35104f9615b8736529223da247f72822d1a869d674cc32b0e3f00 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569040 4872 scope.go:117] "RemoveContainer" containerID="d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569327 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246"} err="failed to get container status \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": rpc error: code = NotFound desc = could not find container \"d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246\": container with ID starting with d33306fb6cbfbc4ac9857a3afa2e52311686f4bc77406493a99fa492e0cfb246 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569353 4872 scope.go:117] "RemoveContainer" containerID="f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569668 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1"} err="failed to get container status \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": rpc error: code = NotFound desc = could not find container \"f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1\": container with ID starting with f524249edf420d6eb85d2b0dbaf162ab7dc6c5b48d9b1a6cf2ebe6927d3fa4d1 not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569696 4872 scope.go:117] "RemoveContainer" containerID="bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.569934 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd"} err="failed to get container status \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": rpc error: code = NotFound desc = could not find container \"bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd\": container with ID starting with bb44c4ccc780a96f5ab534317396507cc6e1b2883f7b26d28cbf1a2a01c9a2fd not found: ID does not exist" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.701092 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.712266 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.722027 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.722547 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-notification-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.722637 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-notification-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.722721 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-central-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.722781 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-central-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.722856 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="sg-core" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.722905 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="sg-core" Oct 09 08:38:24 crc kubenswrapper[4872]: E1009 08:38:24.722970 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="proxy-httpd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.723018 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="proxy-httpd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.723241 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="proxy-httpd" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.723347 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-notification-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.723400 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="sg-core" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.723457 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="53077894-a0c8-488c-9e7b-794ec805163f" containerName="ceilometer-central-agent" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.725246 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.727998 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.728234 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.730139 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.730748 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.901979 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-scripts\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903079 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-log-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903345 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7wcj\" (UniqueName: \"kubernetes.io/projected/d6a77283-bbb4-4a6b-9713-f569e57ed673-kube-api-access-q7wcj\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903427 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-config-data\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903638 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-run-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.903773 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:24 crc kubenswrapper[4872]: I1009 08:38:24.904056 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006106 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-config-data\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006172 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-run-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006205 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006298 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006363 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-scripts\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006402 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-log-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006479 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006699 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-run-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006862 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d6a77283-bbb4-4a6b-9713-f569e57ed673-log-httpd\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.006948 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7wcj\" (UniqueName: \"kubernetes.io/projected/d6a77283-bbb4-4a6b-9713-f569e57ed673-kube-api-access-q7wcj\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.013318 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.014985 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-config-data\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.015042 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-scripts\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.015333 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.015507 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d6a77283-bbb4-4a6b-9713-f569e57ed673-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.025102 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7wcj\" (UniqueName: \"kubernetes.io/projected/d6a77283-bbb4-4a6b-9713-f569e57ed673-kube-api-access-q7wcj\") pod \"ceilometer-0\" (UID: \"d6a77283-bbb4-4a6b-9713-f569e57ed673\") " pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.054984 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.550949 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 09 08:38:25 crc kubenswrapper[4872]: W1009 08:38:25.568074 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6a77283_bbb4_4a6b_9713_f569e57ed673.slice/crio-2a87a522be631ffb848dfb36dc28a87e825c76754c682091d232090cda295636 WatchSource:0}: Error finding container 2a87a522be631ffb848dfb36dc28a87e825c76754c682091d232090cda295636: Status 404 returned error can't find the container with id 2a87a522be631ffb848dfb36dc28a87e825c76754c682091d232090cda295636 Oct 09 08:38:25 crc kubenswrapper[4872]: I1009 08:38:25.932893 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.128423 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data\") pod \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.128484 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs\") pod \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.128554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxxg4\" (UniqueName: \"kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4\") pod \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.128573 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle\") pod \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\" (UID: \"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f\") " Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.130118 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs" (OuterVolumeSpecName: "logs") pod "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" (UID: "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.134471 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4" (OuterVolumeSpecName: "kube-api-access-bxxg4") pod "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" (UID: "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f"). InnerVolumeSpecName "kube-api-access-bxxg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.172748 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" (UID: "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.180835 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data" (OuterVolumeSpecName: "config-data") pod "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" (UID: "3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.230928 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxxg4\" (UniqueName: \"kubernetes.io/projected/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-kube-api-access-bxxg4\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.230979 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.230993 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.231010 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.376399 4872 generic.go:334] "Generic (PLEG): container finished" podID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerID="65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296" exitCode=0 Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.376442 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerDied","Data":"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296"} Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.376511 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.376877 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f","Type":"ContainerDied","Data":"64c83675412d7df7c984c57d0b4320f711b8414a9106ece07c72c11a77d734a6"} Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.376913 4872 scope.go:117] "RemoveContainer" containerID="65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.380136 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6a77283-bbb4-4a6b-9713-f569e57ed673","Type":"ContainerStarted","Data":"b64629ea9d85cd7fcbfb7c8658b726f5bcedc5871a5e47dc7ecbac2e149fb1eb"} Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.380179 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6a77283-bbb4-4a6b-9713-f569e57ed673","Type":"ContainerStarted","Data":"2a87a522be631ffb848dfb36dc28a87e825c76754c682091d232090cda295636"} Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.396229 4872 scope.go:117] "RemoveContainer" containerID="8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.422778 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.439314 4872 scope.go:117] "RemoveContainer" containerID="65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296" Oct 09 08:38:26 crc kubenswrapper[4872]: E1009 08:38:26.443257 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296\": container with ID starting with 65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296 not found: ID does not exist" containerID="65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.443302 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296"} err="failed to get container status \"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296\": rpc error: code = NotFound desc = could not find container \"65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296\": container with ID starting with 65d065294d3b3baf0923b50ca0dcaf33de9af2e2615d2a27c2e857c65fb28296 not found: ID does not exist" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.443335 4872 scope.go:117] "RemoveContainer" containerID="8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae" Oct 09 08:38:26 crc kubenswrapper[4872]: E1009 08:38:26.444832 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae\": container with ID starting with 8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae not found: ID does not exist" containerID="8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.444877 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae"} err="failed to get container status \"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae\": rpc error: code = NotFound desc = could not find container \"8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae\": container with ID starting with 8017813d09638cfa357774194fb02c95ad4203e14604c44e7c834dfd26805fae not found: ID does not exist" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.446224 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.455915 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:26 crc kubenswrapper[4872]: E1009 08:38:26.456457 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-api" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.456479 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-api" Oct 09 08:38:26 crc kubenswrapper[4872]: E1009 08:38:26.456506 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-log" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.456515 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-log" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.456805 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-api" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.456826 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" containerName="nova-api-log" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.458194 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.464721 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.465316 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.465657 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.496986 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f" path="/var/lib/kubelet/pods/3f680e1a-99ae-41b7-b6e0-8d9e22fcee6f/volumes" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.497833 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53077894-a0c8-488c-9e7b-794ec805163f" path="/var/lib/kubelet/pods/53077894-a0c8-488c-9e7b-794ec805163f/volumes" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.498738 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638438 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638621 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638791 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638843 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638882 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.638935 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8bl8\" (UniqueName: \"kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741161 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741226 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741268 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741317 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8bl8\" (UniqueName: \"kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741359 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.741494 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.742770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.746190 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.747067 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.747598 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.747811 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.758294 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8bl8\" (UniqueName: \"kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8\") pod \"nova-api-0\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " pod="openstack/nova-api-0" Oct 09 08:38:26 crc kubenswrapper[4872]: I1009 08:38:26.785664 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:27 crc kubenswrapper[4872]: W1009 08:38:27.259363 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3813eb52_d33e_439e_a221_446e3133ae2c.slice/crio-2e5389ff1bbca4ca43470bf7109cf71c5dba4cd855df98acc1deb0b08a0f0a33 WatchSource:0}: Error finding container 2e5389ff1bbca4ca43470bf7109cf71c5dba4cd855df98acc1deb0b08a0f0a33: Status 404 returned error can't find the container with id 2e5389ff1bbca4ca43470bf7109cf71c5dba4cd855df98acc1deb0b08a0f0a33 Oct 09 08:38:27 crc kubenswrapper[4872]: I1009 08:38:27.259505 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:27 crc kubenswrapper[4872]: I1009 08:38:27.393131 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6a77283-bbb4-4a6b-9713-f569e57ed673","Type":"ContainerStarted","Data":"d0c1df81fba5c3610f1304f9568ac709d9a023ee892cf1c58a0bb6263d71bc11"} Oct 09 08:38:27 crc kubenswrapper[4872]: I1009 08:38:27.395192 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerStarted","Data":"2e5389ff1bbca4ca43470bf7109cf71c5dba4cd855df98acc1deb0b08a0f0a33"} Oct 09 08:38:27 crc kubenswrapper[4872]: I1009 08:38:27.638206 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:27 crc kubenswrapper[4872]: I1009 08:38:27.657769 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.405015 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6a77283-bbb4-4a6b-9713-f569e57ed673","Type":"ContainerStarted","Data":"efef928710f0a8bc8dfd258f44d435e0e64b11f5e94f71ab771be17a2bcd162b"} Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.407028 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerStarted","Data":"9d0442851fd8e77b4a0ba7867e1187dababf03229c93dcef70746a023ec123d8"} Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.407079 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerStarted","Data":"50e08aed1952a4369e6c181898b0f0672133ef12e6d09862888e60b8e84408fd"} Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.423267 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.423242203 podStartE2EDuration="2.423242203s" podCreationTimestamp="2025-10-09 08:38:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:28.422130871 +0000 UTC m=+1266.612659517" watchObservedRunningTime="2025-10-09 08:38:28.423242203 +0000 UTC m=+1266.613770849" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.434123 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.624766 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-5db52"] Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.626277 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.630276 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.630323 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.635831 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-5db52"] Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.778606 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzmv7\" (UniqueName: \"kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.778685 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.779246 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.779322 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.881591 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.881997 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.882107 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzmv7\" (UniqueName: \"kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.882163 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.887289 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.889554 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.898770 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.902347 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzmv7\" (UniqueName: \"kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7\") pod \"nova-cell1-cell-mapping-5db52\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:28 crc kubenswrapper[4872]: I1009 08:38:28.947054 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.422963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d6a77283-bbb4-4a6b-9713-f569e57ed673","Type":"ContainerStarted","Data":"6ec298908e29b91c559a5bb42bdb15d55bf1b9e6abf4c42b1d9489c88777d03d"} Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.463311 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-5db52"] Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.464343 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.250235281 podStartE2EDuration="5.464322874s" podCreationTimestamp="2025-10-09 08:38:24 +0000 UTC" firstStartedPulling="2025-10-09 08:38:25.570980474 +0000 UTC m=+1263.761509100" lastFinishedPulling="2025-10-09 08:38:28.785068067 +0000 UTC m=+1266.975596693" observedRunningTime="2025-10-09 08:38:29.449195177 +0000 UTC m=+1267.639723823" watchObservedRunningTime="2025-10-09 08:38:29.464322874 +0000 UTC m=+1267.654851500" Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.801259 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.860659 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:38:29 crc kubenswrapper[4872]: I1009 08:38:29.860946 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="dnsmasq-dns" containerID="cri-o://f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2" gracePeriod=10 Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.352332 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.437926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-5db52" event={"ID":"90e90359-2f2f-4f14-adf5-43f088b166a4","Type":"ContainerStarted","Data":"0b71167146630c71d28cbeb1eb3662716086ad25dfc4b1391d957e7be1a53cc6"} Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.437979 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-5db52" event={"ID":"90e90359-2f2f-4f14-adf5-43f088b166a4","Type":"ContainerStarted","Data":"639712b9a9c383f23654fe873acc397826f265b5a3deeab03c61d220dea146f2"} Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.440875 4872 generic.go:334] "Generic (PLEG): container finished" podID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerID="f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2" exitCode=0 Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.440928 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.440977 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" event={"ID":"a2e37277-59e1-49b6-a240-660d3416ff4d","Type":"ContainerDied","Data":"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2"} Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.441048 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-865f5d856f-c2c79" event={"ID":"a2e37277-59e1-49b6-a240-660d3416ff4d","Type":"ContainerDied","Data":"5ee87ff791607adf76f8cbf99040e6130efbf537b343c599834da1d59738aace"} Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.441077 4872 scope.go:117] "RemoveContainer" containerID="f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.441423 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445535 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445601 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445704 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l85wj\" (UniqueName: \"kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445755 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445914 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.445954 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.467736 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-5db52" podStartSLOduration=2.467711726 podStartE2EDuration="2.467711726s" podCreationTimestamp="2025-10-09 08:38:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:30.463040492 +0000 UTC m=+1268.653569128" watchObservedRunningTime="2025-10-09 08:38:30.467711726 +0000 UTC m=+1268.658240352" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.476147 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj" (OuterVolumeSpecName: "kube-api-access-l85wj") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "kube-api-access-l85wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.488070 4872 scope.go:117] "RemoveContainer" containerID="f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.506727 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.524082 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config" (OuterVolumeSpecName: "config") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.524433 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.531812 4872 scope.go:117] "RemoveContainer" containerID="f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2" Oct 09 08:38:30 crc kubenswrapper[4872]: E1009 08:38:30.532317 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2\": container with ID starting with f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2 not found: ID does not exist" containerID="f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.532433 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2"} err="failed to get container status \"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2\": rpc error: code = NotFound desc = could not find container \"f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2\": container with ID starting with f70be93d9c0e7523f20454a892287786420f87d4ba75ea16007b4a62833544e2 not found: ID does not exist" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.532530 4872 scope.go:117] "RemoveContainer" containerID="f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0" Oct 09 08:38:30 crc kubenswrapper[4872]: E1009 08:38:30.535525 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0\": container with ID starting with f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0 not found: ID does not exist" containerID="f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.535583 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0"} err="failed to get container status \"f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0\": rpc error: code = NotFound desc = could not find container \"f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0\": container with ID starting with f053f4e9fc690edc32dc43c082f349f786886316a6b6072c0946d49e7492e4a0 not found: ID does not exist" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.540232 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.547774 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.548256 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") pod \"a2e37277-59e1-49b6-a240-660d3416ff4d\" (UID: \"a2e37277-59e1-49b6-a240-660d3416ff4d\") " Oct 09 08:38:30 crc kubenswrapper[4872]: W1009 08:38:30.548374 4872 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/a2e37277-59e1-49b6-a240-660d3416ff4d/volumes/kubernetes.io~configmap/ovsdbserver-nb Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.548393 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2e37277-59e1-49b6-a240-660d3416ff4d" (UID: "a2e37277-59e1-49b6-a240-660d3416ff4d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549196 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549225 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549239 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549251 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549264 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l85wj\" (UniqueName: \"kubernetes.io/projected/a2e37277-59e1-49b6-a240-660d3416ff4d-kube-api-access-l85wj\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.549276 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e37277-59e1-49b6-a240-660d3416ff4d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.794209 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:38:30 crc kubenswrapper[4872]: I1009 08:38:30.801037 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-865f5d856f-c2c79"] Oct 09 08:38:32 crc kubenswrapper[4872]: I1009 08:38:32.474883 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" path="/var/lib/kubelet/pods/a2e37277-59e1-49b6-a240-660d3416ff4d/volumes" Oct 09 08:38:32 crc kubenswrapper[4872]: E1009 08:38:32.480064 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e\": RecentStats: unable to find data in memory cache]" Oct 09 08:38:35 crc kubenswrapper[4872]: I1009 08:38:35.507097 4872 generic.go:334] "Generic (PLEG): container finished" podID="90e90359-2f2f-4f14-adf5-43f088b166a4" containerID="0b71167146630c71d28cbeb1eb3662716086ad25dfc4b1391d957e7be1a53cc6" exitCode=0 Oct 09 08:38:35 crc kubenswrapper[4872]: I1009 08:38:35.507185 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-5db52" event={"ID":"90e90359-2f2f-4f14-adf5-43f088b166a4","Type":"ContainerDied","Data":"0b71167146630c71d28cbeb1eb3662716086ad25dfc4b1391d957e7be1a53cc6"} Oct 09 08:38:36 crc kubenswrapper[4872]: I1009 08:38:36.786959 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:36 crc kubenswrapper[4872]: I1009 08:38:36.787005 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:36 crc kubenswrapper[4872]: I1009 08:38:36.931009 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.007316 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts\") pod \"90e90359-2f2f-4f14-adf5-43f088b166a4\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.007399 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle\") pod \"90e90359-2f2f-4f14-adf5-43f088b166a4\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.007563 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data\") pod \"90e90359-2f2f-4f14-adf5-43f088b166a4\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.007623 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzmv7\" (UniqueName: \"kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7\") pod \"90e90359-2f2f-4f14-adf5-43f088b166a4\" (UID: \"90e90359-2f2f-4f14-adf5-43f088b166a4\") " Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.014951 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7" (OuterVolumeSpecName: "kube-api-access-zzmv7") pod "90e90359-2f2f-4f14-adf5-43f088b166a4" (UID: "90e90359-2f2f-4f14-adf5-43f088b166a4"). InnerVolumeSpecName "kube-api-access-zzmv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.019893 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts" (OuterVolumeSpecName: "scripts") pod "90e90359-2f2f-4f14-adf5-43f088b166a4" (UID: "90e90359-2f2f-4f14-adf5-43f088b166a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.035806 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data" (OuterVolumeSpecName: "config-data") pod "90e90359-2f2f-4f14-adf5-43f088b166a4" (UID: "90e90359-2f2f-4f14-adf5-43f088b166a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.036970 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90e90359-2f2f-4f14-adf5-43f088b166a4" (UID: "90e90359-2f2f-4f14-adf5-43f088b166a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.108897 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.108930 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzmv7\" (UniqueName: \"kubernetes.io/projected/90e90359-2f2f-4f14-adf5-43f088b166a4-kube-api-access-zzmv7\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.108944 4872 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-scripts\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.108955 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90e90359-2f2f-4f14-adf5-43f088b166a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.531963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-5db52" event={"ID":"90e90359-2f2f-4f14-adf5-43f088b166a4","Type":"ContainerDied","Data":"639712b9a9c383f23654fe873acc397826f265b5a3deeab03c61d220dea146f2"} Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.532018 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="639712b9a9c383f23654fe873acc397826f265b5a3deeab03c61d220dea146f2" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.532092 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-5db52" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.796749 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.797098 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-log" containerID="cri-o://50e08aed1952a4369e6c181898b0f0672133ef12e6d09862888e60b8e84408fd" gracePeriod=30 Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.797200 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-api" containerID="cri-o://9d0442851fd8e77b4a0ba7867e1187dababf03229c93dcef70746a023ec123d8" gracePeriod=30 Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.803705 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.803748 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.202:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.810171 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.810452 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="15a338c3-8256-4891-a0b8-ffab1db121a5" containerName="nova-scheduler-scheduler" containerID="cri-o://25523c74e5ffa42f87f4263e193bfdd893ab2dcd520ea6b60443c8b2ecc6fa65" gracePeriod=30 Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.826960 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.827206 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-log" containerID="cri-o://668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3" gracePeriod=30 Oct 09 08:38:37 crc kubenswrapper[4872]: I1009 08:38:37.827339 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-metadata" containerID="cri-o://ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4" gracePeriod=30 Oct 09 08:38:38 crc kubenswrapper[4872]: I1009 08:38:38.567446 4872 generic.go:334] "Generic (PLEG): container finished" podID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerID="668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3" exitCode=143 Oct 09 08:38:38 crc kubenswrapper[4872]: I1009 08:38:38.567543 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerDied","Data":"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3"} Oct 09 08:38:38 crc kubenswrapper[4872]: I1009 08:38:38.574153 4872 generic.go:334] "Generic (PLEG): container finished" podID="3813eb52-d33e-439e-a221-446e3133ae2c" containerID="50e08aed1952a4369e6c181898b0f0672133ef12e6d09862888e60b8e84408fd" exitCode=143 Oct 09 08:38:38 crc kubenswrapper[4872]: I1009 08:38:38.574209 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerDied","Data":"50e08aed1952a4369e6c181898b0f0672133ef12e6d09862888e60b8e84408fd"} Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.490290 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.593532 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle\") pod \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.593585 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc6ff\" (UniqueName: \"kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff\") pod \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.593613 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data\") pod \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.593817 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs\") pod \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.593895 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs\") pod \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\" (UID: \"10f65ca4-ef8b-40de-88eb-e4e42f78e553\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.597694 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs" (OuterVolumeSpecName: "logs") pod "10f65ca4-ef8b-40de-88eb-e4e42f78e553" (UID: "10f65ca4-ef8b-40de-88eb-e4e42f78e553"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608504 4872 generic.go:334] "Generic (PLEG): container finished" podID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerID="ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4" exitCode=0 Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608578 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerDied","Data":"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4"} Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608605 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"10f65ca4-ef8b-40de-88eb-e4e42f78e553","Type":"ContainerDied","Data":"38a262cd0592822345d5737affbae53bb8b937f8bbe242657e1fda8716ef9e87"} Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608622 4872 scope.go:117] "RemoveContainer" containerID="ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608654 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.608772 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff" (OuterVolumeSpecName: "kube-api-access-qc6ff") pod "10f65ca4-ef8b-40de-88eb-e4e42f78e553" (UID: "10f65ca4-ef8b-40de-88eb-e4e42f78e553"). InnerVolumeSpecName "kube-api-access-qc6ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.612353 4872 generic.go:334] "Generic (PLEG): container finished" podID="15a338c3-8256-4891-a0b8-ffab1db121a5" containerID="25523c74e5ffa42f87f4263e193bfdd893ab2dcd520ea6b60443c8b2ecc6fa65" exitCode=0 Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.612376 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a338c3-8256-4891-a0b8-ffab1db121a5","Type":"ContainerDied","Data":"25523c74e5ffa42f87f4263e193bfdd893ab2dcd520ea6b60443c8b2ecc6fa65"} Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.635855 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data" (OuterVolumeSpecName: "config-data") pod "10f65ca4-ef8b-40de-88eb-e4e42f78e553" (UID: "10f65ca4-ef8b-40de-88eb-e4e42f78e553"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.647725 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10f65ca4-ef8b-40de-88eb-e4e42f78e553" (UID: "10f65ca4-ef8b-40de-88eb-e4e42f78e553"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.651899 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "10f65ca4-ef8b-40de-88eb-e4e42f78e553" (UID: "10f65ca4-ef8b-40de-88eb-e4e42f78e553"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.696527 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10f65ca4-ef8b-40de-88eb-e4e42f78e553-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.696576 4872 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.696592 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.696604 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc6ff\" (UniqueName: \"kubernetes.io/projected/10f65ca4-ef8b-40de-88eb-e4e42f78e553-kube-api-access-qc6ff\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.696615 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10f65ca4-ef8b-40de-88eb-e4e42f78e553-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.724294 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.728023 4872 scope.go:117] "RemoveContainer" containerID="668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.754958 4872 scope.go:117] "RemoveContainer" containerID="ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.755461 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4\": container with ID starting with ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4 not found: ID does not exist" containerID="ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.755498 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4"} err="failed to get container status \"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4\": rpc error: code = NotFound desc = could not find container \"ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4\": container with ID starting with ef299da6cdde78cd8aa4223be07ae1f68c52f05fe46d2c15dab237b84a8a34b4 not found: ID does not exist" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.755524 4872 scope.go:117] "RemoveContainer" containerID="668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.755998 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3\": container with ID starting with 668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3 not found: ID does not exist" containerID="668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.756046 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3"} err="failed to get container status \"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3\": rpc error: code = NotFound desc = could not find container \"668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3\": container with ID starting with 668b5f40067b90878f2fdd5fe2fb4f22b8723b8536f4ba5b6a169e0dcbd4c0b3 not found: ID does not exist" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.899588 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle\") pod \"15a338c3-8256-4891-a0b8-ffab1db121a5\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.899942 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data\") pod \"15a338c3-8256-4891-a0b8-ffab1db121a5\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.900073 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xqlb\" (UniqueName: \"kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb\") pod \"15a338c3-8256-4891-a0b8-ffab1db121a5\" (UID: \"15a338c3-8256-4891-a0b8-ffab1db121a5\") " Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.903332 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb" (OuterVolumeSpecName: "kube-api-access-5xqlb") pod "15a338c3-8256-4891-a0b8-ffab1db121a5" (UID: "15a338c3-8256-4891-a0b8-ffab1db121a5"). InnerVolumeSpecName "kube-api-access-5xqlb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.926956 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data" (OuterVolumeSpecName: "config-data") pod "15a338c3-8256-4891-a0b8-ffab1db121a5" (UID: "15a338c3-8256-4891-a0b8-ffab1db121a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.936230 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "15a338c3-8256-4891-a0b8-ffab1db121a5" (UID: "15a338c3-8256-4891-a0b8-ffab1db121a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.944801 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.952796 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.972856 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973361 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-metadata" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973391 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-metadata" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973425 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a338c3-8256-4891-a0b8-ffab1db121a5" containerName="nova-scheduler-scheduler" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973434 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a338c3-8256-4891-a0b8-ffab1db121a5" containerName="nova-scheduler-scheduler" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973460 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="init" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973468 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="init" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973488 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e90359-2f2f-4f14-adf5-43f088b166a4" containerName="nova-manage" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973497 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e90359-2f2f-4f14-adf5-43f088b166a4" containerName="nova-manage" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973517 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-log" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973525 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-log" Oct 09 08:38:41 crc kubenswrapper[4872]: E1009 08:38:41.973545 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="dnsmasq-dns" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973554 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="dnsmasq-dns" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973786 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e37277-59e1-49b6-a240-660d3416ff4d" containerName="dnsmasq-dns" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973820 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e90359-2f2f-4f14-adf5-43f088b166a4" containerName="nova-manage" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973845 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-metadata" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973871 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" containerName="nova-metadata-log" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.973889 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a338c3-8256-4891-a0b8-ffab1db121a5" containerName="nova-scheduler-scheduler" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.975160 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.977517 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.977915 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 09 08:38:41 crc kubenswrapper[4872]: I1009 08:38:41.982946 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.003242 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.003276 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xqlb\" (UniqueName: \"kubernetes.io/projected/15a338c3-8256-4891-a0b8-ffab1db121a5-kube-api-access-5xqlb\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.003285 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15a338c3-8256-4891-a0b8-ffab1db121a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.104792 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-logs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.104925 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-config-data\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.104962 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.105047 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jsn72\" (UniqueName: \"kubernetes.io/projected/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-kube-api-access-jsn72\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.105215 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.206918 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.207033 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-logs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.207529 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-logs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.207612 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-config-data\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.208048 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.208104 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jsn72\" (UniqueName: \"kubernetes.io/projected/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-kube-api-access-jsn72\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.211243 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.211414 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-config-data\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.211432 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.224134 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jsn72\" (UniqueName: \"kubernetes.io/projected/6f6917d5-6f1e-4f7f-b651-14ee02e1582c-kube-api-access-jsn72\") pod \"nova-metadata-0\" (UID: \"6f6917d5-6f1e-4f7f-b651-14ee02e1582c\") " pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.296018 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.493695 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10f65ca4-ef8b-40de-88eb-e4e42f78e553" path="/var/lib/kubelet/pods/10f65ca4-ef8b-40de-88eb-e4e42f78e553/volumes" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.622820 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"15a338c3-8256-4891-a0b8-ffab1db121a5","Type":"ContainerDied","Data":"a95d26dc58438386fa7fb27fddfa97bf5a1237eeee353abc39839a4685946c71"} Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.622868 4872 scope.go:117] "RemoveContainer" containerID="25523c74e5ffa42f87f4263e193bfdd893ab2dcd520ea6b60443c8b2ecc6fa65" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.622892 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.648576 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.670931 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.680974 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.682662 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.688939 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.690062 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 09 08:38:42 crc kubenswrapper[4872]: W1009 08:38:42.765434 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f6917d5_6f1e_4f7f_b651_14ee02e1582c.slice/crio-3ec81f45fff2ef5fb8dc747a73705b46ef11922a00d7ff93b2daf3b907bbd1f2 WatchSource:0}: Error finding container 3ec81f45fff2ef5fb8dc747a73705b46ef11922a00d7ff93b2daf3b907bbd1f2: Status 404 returned error can't find the container with id 3ec81f45fff2ef5fb8dc747a73705b46ef11922a00d7ff93b2daf3b907bbd1f2 Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.769708 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 09 08:38:42 crc kubenswrapper[4872]: E1009 08:38:42.779698 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e\": RecentStats: unable to find data in memory cache]" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.820680 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-config-data\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.820795 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.820867 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzp8r\" (UniqueName: \"kubernetes.io/projected/605ee569-e22f-4c02-8a31-34a9a92d7497-kube-api-access-kzp8r\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.922460 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-config-data\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.922540 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.922580 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzp8r\" (UniqueName: \"kubernetes.io/projected/605ee569-e22f-4c02-8a31-34a9a92d7497-kube-api-access-kzp8r\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.928792 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-config-data\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.929818 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/605ee569-e22f-4c02-8a31-34a9a92d7497-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:42 crc kubenswrapper[4872]: I1009 08:38:42.941915 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzp8r\" (UniqueName: \"kubernetes.io/projected/605ee569-e22f-4c02-8a31-34a9a92d7497-kube-api-access-kzp8r\") pod \"nova-scheduler-0\" (UID: \"605ee569-e22f-4c02-8a31-34a9a92d7497\") " pod="openstack/nova-scheduler-0" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.010416 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 09 08:38:43 crc kubenswrapper[4872]: W1009 08:38:43.461971 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod605ee569_e22f_4c02_8a31_34a9a92d7497.slice/crio-5224fa4ed0179654c0de1aea28c13c5c671c66bb3581e30617e0529fddf841f5 WatchSource:0}: Error finding container 5224fa4ed0179654c0de1aea28c13c5c671c66bb3581e30617e0529fddf841f5: Status 404 returned error can't find the container with id 5224fa4ed0179654c0de1aea28c13c5c671c66bb3581e30617e0529fddf841f5 Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.465628 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.662260 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f6917d5-6f1e-4f7f-b651-14ee02e1582c","Type":"ContainerStarted","Data":"5a0615adec6875dca1553ba4e420595ef5f234bb2d6e5c89fc1ee204a4191348"} Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.662796 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f6917d5-6f1e-4f7f-b651-14ee02e1582c","Type":"ContainerStarted","Data":"ea03789ce4186a1a653ebebb297b91ff83f375c78b3f384d8995b2bd788c65ea"} Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.662813 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"6f6917d5-6f1e-4f7f-b651-14ee02e1582c","Type":"ContainerStarted","Data":"3ec81f45fff2ef5fb8dc747a73705b46ef11922a00d7ff93b2daf3b907bbd1f2"} Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.665542 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"605ee569-e22f-4c02-8a31-34a9a92d7497","Type":"ContainerStarted","Data":"5224fa4ed0179654c0de1aea28c13c5c671c66bb3581e30617e0529fddf841f5"} Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.668696 4872 generic.go:334] "Generic (PLEG): container finished" podID="3813eb52-d33e-439e-a221-446e3133ae2c" containerID="9d0442851fd8e77b4a0ba7867e1187dababf03229c93dcef70746a023ec123d8" exitCode=0 Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.668722 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerDied","Data":"9d0442851fd8e77b4a0ba7867e1187dababf03229c93dcef70746a023ec123d8"} Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.690002 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.689984034 podStartE2EDuration="2.689984034s" podCreationTimestamp="2025-10-09 08:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:43.682793117 +0000 UTC m=+1281.873321743" watchObservedRunningTime="2025-10-09 08:38:43.689984034 +0000 UTC m=+1281.880512660" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.741976 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826008 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826069 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826144 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826232 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826253 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8bl8\" (UniqueName: \"kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826281 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data\") pod \"3813eb52-d33e-439e-a221-446e3133ae2c\" (UID: \"3813eb52-d33e-439e-a221-446e3133ae2c\") " Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.826654 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs" (OuterVolumeSpecName: "logs") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.830511 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8" (OuterVolumeSpecName: "kube-api-access-g8bl8") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "kube-api-access-g8bl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.854745 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.856557 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data" (OuterVolumeSpecName: "config-data") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.877215 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.879812 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3813eb52-d33e-439e-a221-446e3133ae2c" (UID: "3813eb52-d33e-439e-a221-446e3133ae2c"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928621 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928682 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8bl8\" (UniqueName: \"kubernetes.io/projected/3813eb52-d33e-439e-a221-446e3133ae2c-kube-api-access-g8bl8\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928697 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928707 4872 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928719 4872 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3813eb52-d33e-439e-a221-446e3133ae2c-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:43 crc kubenswrapper[4872]: I1009 08:38:43.928729 4872 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3813eb52-d33e-439e-a221-446e3133ae2c-logs\") on node \"crc\" DevicePath \"\"" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.477417 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a338c3-8256-4891-a0b8-ffab1db121a5" path="/var/lib/kubelet/pods/15a338c3-8256-4891-a0b8-ffab1db121a5/volumes" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.684180 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"605ee569-e22f-4c02-8a31-34a9a92d7497","Type":"ContainerStarted","Data":"968805b6caf40db783f72235cf30ccc78e1b556faf8c199bab2e2e576a61ac11"} Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.702412 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3813eb52-d33e-439e-a221-446e3133ae2c","Type":"ContainerDied","Data":"2e5389ff1bbca4ca43470bf7109cf71c5dba4cd855df98acc1deb0b08a0f0a33"} Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.702511 4872 scope.go:117] "RemoveContainer" containerID="9d0442851fd8e77b4a0ba7867e1187dababf03229c93dcef70746a023ec123d8" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.702650 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.719134 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.719107837 podStartE2EDuration="2.719107837s" podCreationTimestamp="2025-10-09 08:38:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:44.702378025 +0000 UTC m=+1282.892906671" watchObservedRunningTime="2025-10-09 08:38:44.719107837 +0000 UTC m=+1282.909636483" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.739849 4872 scope.go:117] "RemoveContainer" containerID="50e08aed1952a4369e6c181898b0f0672133ef12e6d09862888e60b8e84408fd" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.744020 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.756471 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.777922 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:44 crc kubenswrapper[4872]: E1009 08:38:44.778653 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-api" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.779818 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-api" Oct 09 08:38:44 crc kubenswrapper[4872]: E1009 08:38:44.779949 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-log" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.780025 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-log" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.780371 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-api" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.780482 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" containerName="nova-api-log" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.781849 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.793331 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.793507 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.793677 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.807400 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.953532 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-logs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.953593 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.953626 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5sj7\" (UniqueName: \"kubernetes.io/projected/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-kube-api-access-w5sj7\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.954346 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.954543 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:44 crc kubenswrapper[4872]: I1009 08:38:44.954592 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-config-data\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058624 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058733 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058754 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-config-data\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058823 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-logs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058859 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.058897 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5sj7\" (UniqueName: \"kubernetes.io/projected/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-kube-api-access-w5sj7\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.060081 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-logs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.065290 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.077536 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-config-data\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.077676 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.078985 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-public-tls-certs\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.080914 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5sj7\" (UniqueName: \"kubernetes.io/projected/7a95ce78-4b3a-4aea-8351-c39dc3836f4c-kube-api-access-w5sj7\") pod \"nova-api-0\" (UID: \"7a95ce78-4b3a-4aea-8351-c39dc3836f4c\") " pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.110146 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.583488 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 09 08:38:45 crc kubenswrapper[4872]: I1009 08:38:45.717046 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a95ce78-4b3a-4aea-8351-c39dc3836f4c","Type":"ContainerStarted","Data":"503400f32a2731774ee24e1eb9b4213e1ee1502151665b1c6a362e573c173585"} Oct 09 08:38:46 crc kubenswrapper[4872]: I1009 08:38:46.471094 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3813eb52-d33e-439e-a221-446e3133ae2c" path="/var/lib/kubelet/pods/3813eb52-d33e-439e-a221-446e3133ae2c/volumes" Oct 09 08:38:46 crc kubenswrapper[4872]: I1009 08:38:46.728287 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a95ce78-4b3a-4aea-8351-c39dc3836f4c","Type":"ContainerStarted","Data":"316af85d3b3f0457602eb38992ed1baedf9f22917963be10a99210467c22bbd0"} Oct 09 08:38:46 crc kubenswrapper[4872]: I1009 08:38:46.728681 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7a95ce78-4b3a-4aea-8351-c39dc3836f4c","Type":"ContainerStarted","Data":"96c72873948db66a9e7c78920fe5d007b167ed8faab514161112ec717df77e71"} Oct 09 08:38:46 crc kubenswrapper[4872]: I1009 08:38:46.753901 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.75387804 podStartE2EDuration="2.75387804s" podCreationTimestamp="2025-10-09 08:38:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:38:46.747413373 +0000 UTC m=+1284.937942029" watchObservedRunningTime="2025-10-09 08:38:46.75387804 +0000 UTC m=+1284.944406676" Oct 09 08:38:47 crc kubenswrapper[4872]: I1009 08:38:47.296655 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 08:38:47 crc kubenswrapper[4872]: I1009 08:38:47.296697 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 09 08:38:48 crc kubenswrapper[4872]: I1009 08:38:48.011180 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 09 08:38:52 crc kubenswrapper[4872]: I1009 08:38:52.296254 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 08:38:52 crc kubenswrapper[4872]: I1009 08:38:52.296709 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 09 08:38:53 crc kubenswrapper[4872]: I1009 08:38:53.011062 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 09 08:38:53 crc kubenswrapper[4872]: I1009 08:38:53.038460 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 09 08:38:53 crc kubenswrapper[4872]: E1009 08:38:53.105047 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e\": RecentStats: unable to find data in memory cache]" Oct 09 08:38:53 crc kubenswrapper[4872]: I1009 08:38:53.308839 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f6917d5-6f1e-4f7f-b651-14ee02e1582c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:53 crc kubenswrapper[4872]: I1009 08:38:53.308894 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="6f6917d5-6f1e-4f7f-b651-14ee02e1582c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.204:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:53 crc kubenswrapper[4872]: I1009 08:38:53.833952 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 09 08:38:55 crc kubenswrapper[4872]: I1009 08:38:55.064538 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 09 08:38:55 crc kubenswrapper[4872]: I1009 08:38:55.111176 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:55 crc kubenswrapper[4872]: I1009 08:38:55.111217 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 09 08:38:56 crc kubenswrapper[4872]: I1009 08:38:56.122832 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a95ce78-4b3a-4aea-8351-c39dc3836f4c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:38:56 crc kubenswrapper[4872]: I1009 08:38:56.122842 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7a95ce78-4b3a-4aea-8351-c39dc3836f4c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.206:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 09 08:39:02 crc kubenswrapper[4872]: I1009 08:39:02.304443 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 08:39:02 crc kubenswrapper[4872]: I1009 08:39:02.305301 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 09 08:39:02 crc kubenswrapper[4872]: I1009 08:39:02.310429 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 08:39:02 crc kubenswrapper[4872]: I1009 08:39:02.909705 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 09 08:39:03 crc kubenswrapper[4872]: E1009 08:39:03.353694 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e\": RecentStats: unable to find data in memory cache]" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.117801 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.118773 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.119317 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.124298 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.931152 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 09 08:39:05 crc kubenswrapper[4872]: I1009 08:39:05.942041 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 09 08:39:08 crc kubenswrapper[4872]: I1009 08:39:08.157188 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:39:08 crc kubenswrapper[4872]: I1009 08:39:08.157558 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:39:13 crc kubenswrapper[4872]: E1009 08:39:13.641893 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53077894_a0c8_488c_9e7b_794ec805163f.slice/crio-55051598cd4a6bb3993d7ac80be282b2f89cfb9dde5775aa23df9c95ae5c276e\": RecentStats: unable to find data in memory cache]" Oct 09 08:39:14 crc kubenswrapper[4872]: I1009 08:39:14.772589 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:16 crc kubenswrapper[4872]: I1009 08:39:16.034296 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:18 crc kubenswrapper[4872]: I1009 08:39:18.528313 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="rabbitmq" containerID="cri-o://e8fa4656d598511564adf0e133ae4d8bb506b76e5cfa04100134020a2b209312" gracePeriod=604797 Oct 09 08:39:19 crc kubenswrapper[4872]: I1009 08:39:19.993817 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="rabbitmq" containerID="cri-o://9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043" gracePeriod=604797 Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.110463 4872 generic.go:334] "Generic (PLEG): container finished" podID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerID="e8fa4656d598511564adf0e133ae4d8bb506b76e5cfa04100134020a2b209312" exitCode=0 Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.110621 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerDied","Data":"e8fa4656d598511564adf0e133ae4d8bb506b76e5cfa04100134020a2b209312"} Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.111096 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"270c0527-0d26-4738-87a6-6d65e143d7ad","Type":"ContainerDied","Data":"2ff0970f91397e74616bad4cd083744ea93b85359ba2daed45c8b16cecdb813f"} Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.111113 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ff0970f91397e74616bad4cd083744ea93b85359ba2daed45c8b16cecdb813f" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.142355 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268558 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268620 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268660 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268724 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268752 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268770 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268876 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268919 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lv6m9\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268939 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.268986 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.269017 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf\") pod \"270c0527-0d26-4738-87a6-6d65e143d7ad\" (UID: \"270c0527-0d26-4738-87a6-6d65e143d7ad\") " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.269192 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.269375 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.269565 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.269887 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.275725 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "persistence") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.275720 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.276339 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info" (OuterVolumeSpecName: "pod-info") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.276524 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.277024 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9" (OuterVolumeSpecName: "kube-api-access-lv6m9") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "kube-api-access-lv6m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.298789 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data" (OuterVolumeSpecName: "config-data") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.329914 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf" (OuterVolumeSpecName: "server-conf") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371071 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lv6m9\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-kube-api-access-lv6m9\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371107 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371115 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371124 4872 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371152 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371161 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371169 4872 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/270c0527-0d26-4738-87a6-6d65e143d7ad-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371179 4872 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/270c0527-0d26-4738-87a6-6d65e143d7ad-pod-info\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.371187 4872 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/270c0527-0d26-4738-87a6-6d65e143d7ad-server-conf\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.393889 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.397845 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "270c0527-0d26-4738-87a6-6d65e143d7ad" (UID: "270c0527-0d26-4738-87a6-6d65e143d7ad"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.472274 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/270c0527-0d26-4738-87a6-6d65e143d7ad-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:25 crc kubenswrapper[4872]: I1009 08:39:25.472549 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.120045 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.206324 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.233460 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.253456 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:26 crc kubenswrapper[4872]: E1009 08:39:26.254760 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="setup-container" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.254788 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="setup-container" Oct 09 08:39:26 crc kubenswrapper[4872]: E1009 08:39:26.254846 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="rabbitmq" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.254856 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="rabbitmq" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.255229 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" containerName="rabbitmq" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.257584 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.263426 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.264506 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.264557 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.264693 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.264752 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.264851 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.265631 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.281847 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-d6zqp" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398083 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-config-data\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398230 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398393 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw76d\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-kube-api-access-zw76d\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398434 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b180311b-aa96-4a46-bbb2-5c32d18b0593-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398450 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398530 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398547 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b180311b-aa96-4a46-bbb2-5c32d18b0593-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.398684 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.473404 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270c0527-0d26-4738-87a6-6d65e143d7ad" path="/var/lib/kubelet/pods/270c0527-0d26-4738-87a6-6d65e143d7ad/volumes" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501094 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501165 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw76d\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-kube-api-access-zw76d\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501220 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b180311b-aa96-4a46-bbb2-5c32d18b0593-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501239 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501301 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501323 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b180311b-aa96-4a46-bbb2-5c32d18b0593-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501358 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501385 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501475 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-config-data\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.501790 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.502003 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.502766 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.504197 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.504774 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.504999 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b180311b-aa96-4a46-bbb2-5c32d18b0593-config-data\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.508352 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b180311b-aa96-4a46-bbb2-5c32d18b0593-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.512443 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.513721 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b180311b-aa96-4a46-bbb2-5c32d18b0593-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.515128 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.518261 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw76d\" (UniqueName: \"kubernetes.io/projected/b180311b-aa96-4a46-bbb2-5c32d18b0593-kube-api-access-zw76d\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.546484 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"rabbitmq-server-0\" (UID: \"b180311b-aa96-4a46-bbb2-5c32d18b0593\") " pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.603710 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.625631 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.705024 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.705942 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706018 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706067 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706115 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706216 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706259 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706303 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2s5s\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706346 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706387 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.706417 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls\") pod \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\" (UID: \"ecf79d03-8c1e-4c55-87d6-40f02a70be53\") " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.707056 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.707808 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.708048 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.712026 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.724194 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s" (OuterVolumeSpecName: "kube-api-access-q2s5s") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "kube-api-access-q2s5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.725953 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.728801 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "persistence") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.736771 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info" (OuterVolumeSpecName: "pod-info") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.753032 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data" (OuterVolumeSpecName: "config-data") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.786421 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf" (OuterVolumeSpecName: "server-conf") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808218 4872 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-server-conf\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808260 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808272 4872 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ecf79d03-8c1e-4c55-87d6-40f02a70be53-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808282 4872 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808294 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ecf79d03-8c1e-4c55-87d6-40f02a70be53-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808330 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808348 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2s5s\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-kube-api-access-q2s5s\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808359 4872 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ecf79d03-8c1e-4c55-87d6-40f02a70be53-pod-info\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808368 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.808380 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.836331 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.875712 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "ecf79d03-8c1e-4c55-87d6-40f02a70be53" (UID: "ecf79d03-8c1e-4c55-87d6-40f02a70be53"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.910321 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:26 crc kubenswrapper[4872]: I1009 08:39:26.910355 4872 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ecf79d03-8c1e-4c55-87d6-40f02a70be53-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.132904 4872 generic.go:334] "Generic (PLEG): container finished" podID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerID="9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043" exitCode=0 Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.132959 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerDied","Data":"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043"} Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.132994 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"ecf79d03-8c1e-4c55-87d6-40f02a70be53","Type":"ContainerDied","Data":"67fe2f33135b08059728d92a02a8bc349605221425aa5ca3c7aed92eb39f8969"} Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.133033 4872 scope.go:117] "RemoveContainer" containerID="9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.133236 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.163272 4872 scope.go:117] "RemoveContainer" containerID="c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.181189 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.190476 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.206964 4872 scope.go:117] "RemoveContainer" containerID="9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043" Oct 09 08:39:27 crc kubenswrapper[4872]: E1009 08:39:27.207409 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043\": container with ID starting with 9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043 not found: ID does not exist" containerID="9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.207458 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043"} err="failed to get container status \"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043\": rpc error: code = NotFound desc = could not find container \"9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043\": container with ID starting with 9b2a58a07402c37265b3eef301e864b08c686e1c3c50b7bce604c6851a280043 not found: ID does not exist" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.207490 4872 scope.go:117] "RemoveContainer" containerID="c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c" Oct 09 08:39:27 crc kubenswrapper[4872]: E1009 08:39:27.209389 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c\": container with ID starting with c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c not found: ID does not exist" containerID="c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.209428 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c"} err="failed to get container status \"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c\": rpc error: code = NotFound desc = could not find container \"c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c\": container with ID starting with c97518a57b0cee8f736bd6bb8ba5fc54e1a3044e86c8ee5ea7008e84ffc7d89c not found: ID does not exist" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.217124 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.227141 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:27 crc kubenswrapper[4872]: E1009 08:39:27.227665 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="rabbitmq" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.227708 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="rabbitmq" Oct 09 08:39:27 crc kubenswrapper[4872]: E1009 08:39:27.227742 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="setup-container" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.227751 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="setup-container" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.229144 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" containerName="rabbitmq" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.230349 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.240040 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.240234 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.240390 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.240525 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.242975 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.246406 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.248404 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-hm27g" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.265080 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:27 crc kubenswrapper[4872]: W1009 08:39:27.267743 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb180311b_aa96_4a46_bbb2_5c32d18b0593.slice/crio-403d268141fd4570e68bf58c9d154d891c53d6de3dd5d60ab7f06bfef491ce7d WatchSource:0}: Error finding container 403d268141fd4570e68bf58c9d154d891c53d6de3dd5d60ab7f06bfef491ce7d: Status 404 returned error can't find the container with id 403d268141fd4570e68bf58c9d154d891c53d6de3dd5d60ab7f06bfef491ce7d Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341065 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341129 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341152 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341198 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341270 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341308 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341338 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341358 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w94l8\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-kube-api-access-w94l8\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341375 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.341420 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443124 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443189 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443266 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443322 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443340 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443375 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443409 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443434 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w94l8\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-kube-api-access-w94l8\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443454 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443495 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.443822 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.444458 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.445872 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.446382 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.447075 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.453329 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.454021 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.455036 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.455374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.456356 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.467177 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w94l8\" (UniqueName: \"kubernetes.io/projected/06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc-kube-api-access-w94l8\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.478989 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc\") " pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:27 crc kubenswrapper[4872]: I1009 08:39:27.622883 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.058832 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 09 08:39:28 crc kubenswrapper[4872]: W1009 08:39:28.067356 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06d9f9e2_b9c5_4a08_b88c_c94e66a09fcc.slice/crio-5a4457250d68c259eb1c525618b83d9235c534c6b31920e9c01903e94b20bc02 WatchSource:0}: Error finding container 5a4457250d68c259eb1c525618b83d9235c534c6b31920e9c01903e94b20bc02: Status 404 returned error can't find the container with id 5a4457250d68c259eb1c525618b83d9235c534c6b31920e9c01903e94b20bc02 Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.141947 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b180311b-aa96-4a46-bbb2-5c32d18b0593","Type":"ContainerStarted","Data":"403d268141fd4570e68bf58c9d154d891c53d6de3dd5d60ab7f06bfef491ce7d"} Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.144107 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc","Type":"ContainerStarted","Data":"5a4457250d68c259eb1c525618b83d9235c534c6b31920e9c01903e94b20bc02"} Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.475422 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecf79d03-8c1e-4c55-87d6-40f02a70be53" path="/var/lib/kubelet/pods/ecf79d03-8c1e-4c55-87d6-40f02a70be53/volumes" Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.974204 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.976221 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.980571 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Oct 09 08:39:28 crc kubenswrapper[4872]: I1009 08:39:28.990923 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.076823 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.076879 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.076913 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.076942 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9jvc\" (UniqueName: \"kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.077041 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.077377 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.077536 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.179812 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.179877 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.179909 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.179936 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9jvc\" (UniqueName: \"kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.179972 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.180074 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.180123 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.181891 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.181965 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.182082 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.182422 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.182517 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.182610 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.214934 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9jvc\" (UniqueName: \"kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc\") pod \"dnsmasq-dns-5576978c7c-msgzg\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.302242 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:29 crc kubenswrapper[4872]: I1009 08:39:29.807660 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:30 crc kubenswrapper[4872]: I1009 08:39:30.162696 4872 generic.go:334] "Generic (PLEG): container finished" podID="aa161414-76f8-44c4-af75-c2ca575658aa" containerID="c7110c0588155b4a859531ff90cdb9135da5fb43cbd6f96f5902fee1c34f1fbe" exitCode=0 Oct 09 08:39:30 crc kubenswrapper[4872]: I1009 08:39:30.162760 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" event={"ID":"aa161414-76f8-44c4-af75-c2ca575658aa","Type":"ContainerDied","Data":"c7110c0588155b4a859531ff90cdb9135da5fb43cbd6f96f5902fee1c34f1fbe"} Oct 09 08:39:30 crc kubenswrapper[4872]: I1009 08:39:30.162783 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" event={"ID":"aa161414-76f8-44c4-af75-c2ca575658aa","Type":"ContainerStarted","Data":"8b1ba23828c5ac6a2d5c658e217a41d26522887e41be5a39cf810f08a7774f28"} Oct 09 08:39:30 crc kubenswrapper[4872]: I1009 08:39:30.165062 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b180311b-aa96-4a46-bbb2-5c32d18b0593","Type":"ContainerStarted","Data":"a02c9dbf41d7104220acf8379aba57674f2817d318c4b709143654a79f7b485b"} Oct 09 08:39:30 crc kubenswrapper[4872]: I1009 08:39:30.167444 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc","Type":"ContainerStarted","Data":"5d5262403109bedd850dfc8654cdb60ad8e8a618174df50bc978500e60a58aca"} Oct 09 08:39:31 crc kubenswrapper[4872]: I1009 08:39:31.178966 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" event={"ID":"aa161414-76f8-44c4-af75-c2ca575658aa","Type":"ContainerStarted","Data":"681e731370014a39b6a07c191b12504f3a40e0d0ff39f36f69cd2a0c1fecfc7c"} Oct 09 08:39:32 crc kubenswrapper[4872]: I1009 08:39:32.188509 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:38 crc kubenswrapper[4872]: I1009 08:39:38.157427 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:39:38 crc kubenswrapper[4872]: I1009 08:39:38.159018 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.304549 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.341727 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" podStartSLOduration=11.3417043 podStartE2EDuration="11.3417043s" podCreationTimestamp="2025-10-09 08:39:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:39:31.201092371 +0000 UTC m=+1329.391621007" watchObservedRunningTime="2025-10-09 08:39:39.3417043 +0000 UTC m=+1337.532232946" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.389295 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.389701 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="dnsmasq-dns" containerID="cri-o://0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e" gracePeriod=10 Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.515211 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-bd65m"] Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.517922 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.545113 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-bd65m"] Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595006 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595097 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595119 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595133 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595232 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9s4rs\" (UniqueName: \"kubernetes.io/projected/1d643d17-4e67-41f2-a8fc-f9790033d364-kube-api-access-9s4rs\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595384 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.595502 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-config\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697362 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697386 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697401 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697418 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9s4rs\" (UniqueName: \"kubernetes.io/projected/1d643d17-4e67-41f2-a8fc-f9790033d364-kube-api-access-9s4rs\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697670 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.697716 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-config\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.698732 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-sb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.698755 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-config\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.698749 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-ovsdbserver-nb\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.698786 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-openstack-edpm-ipam\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.699338 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-svc\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.699998 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1d643d17-4e67-41f2-a8fc-f9790033d364-dns-swift-storage-0\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.727747 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9s4rs\" (UniqueName: \"kubernetes.io/projected/1d643d17-4e67-41f2-a8fc-f9790033d364-kube-api-access-9s4rs\") pod \"dnsmasq-dns-8c6f6df99-bd65m\" (UID: \"1d643d17-4e67-41f2-a8fc-f9790033d364\") " pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.872237 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:39 crc kubenswrapper[4872]: I1009 08:39:39.994535 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.104612 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shl2j\" (UniqueName: \"kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.104697 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.105529 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.105615 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.108137 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.108274 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc\") pod \"a65d363b-6367-4f29-9d54-9270f6122112\" (UID: \"a65d363b-6367-4f29-9d54-9270f6122112\") " Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.111816 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j" (OuterVolumeSpecName: "kube-api-access-shl2j") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "kube-api-access-shl2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.166079 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.166292 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.169275 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.170192 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config" (OuterVolumeSpecName: "config") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.170712 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a65d363b-6367-4f29-9d54-9270f6122112" (UID: "a65d363b-6367-4f29-9d54-9270f6122112"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210649 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210677 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210687 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shl2j\" (UniqueName: \"kubernetes.io/projected/a65d363b-6367-4f29-9d54-9270f6122112-kube-api-access-shl2j\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210697 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210707 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.210716 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a65d363b-6367-4f29-9d54-9270f6122112-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.268108 4872 generic.go:334] "Generic (PLEG): container finished" podID="a65d363b-6367-4f29-9d54-9270f6122112" containerID="0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e" exitCode=0 Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.268149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" event={"ID":"a65d363b-6367-4f29-9d54-9270f6122112","Type":"ContainerDied","Data":"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e"} Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.268166 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.268185 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" event={"ID":"a65d363b-6367-4f29-9d54-9270f6122112","Type":"ContainerDied","Data":"22bf5d6b2c61e3ec660057366f6465f2b8916352768bc69856ff50c2520ce637"} Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.268216 4872 scope.go:117] "RemoveContainer" containerID="0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.291162 4872 scope.go:117] "RemoveContainer" containerID="8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.309769 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.317544 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c7b6c5df9-d4k6h"] Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.320601 4872 scope.go:117] "RemoveContainer" containerID="0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e" Oct 09 08:39:40 crc kubenswrapper[4872]: E1009 08:39:40.321206 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e\": container with ID starting with 0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e not found: ID does not exist" containerID="0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.321244 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e"} err="failed to get container status \"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e\": rpc error: code = NotFound desc = could not find container \"0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e\": container with ID starting with 0eb5c24fa992f4e1f5a8ede09e879c8de6c21de7a28a69397de498c1e436909e not found: ID does not exist" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.321271 4872 scope.go:117] "RemoveContainer" containerID="8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d" Oct 09 08:39:40 crc kubenswrapper[4872]: E1009 08:39:40.321716 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d\": container with ID starting with 8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d not found: ID does not exist" containerID="8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.321767 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d"} err="failed to get container status \"8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d\": rpc error: code = NotFound desc = could not find container \"8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d\": container with ID starting with 8cb46b1c3f4403ab8d78d765fdc7d4c99af3f495797c12582651b4ea3f536a6d not found: ID does not exist" Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.349976 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8c6f6df99-bd65m"] Oct 09 08:39:40 crc kubenswrapper[4872]: W1009 08:39:40.352776 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d643d17_4e67_41f2_a8fc_f9790033d364.slice/crio-fb0bf155d8f663315e044297efdfd5f7d9454d805c913d69f225598d90571a6f WatchSource:0}: Error finding container fb0bf155d8f663315e044297efdfd5f7d9454d805c913d69f225598d90571a6f: Status 404 returned error can't find the container with id fb0bf155d8f663315e044297efdfd5f7d9454d805c913d69f225598d90571a6f Oct 09 08:39:40 crc kubenswrapper[4872]: I1009 08:39:40.477344 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a65d363b-6367-4f29-9d54-9270f6122112" path="/var/lib/kubelet/pods/a65d363b-6367-4f29-9d54-9270f6122112/volumes" Oct 09 08:39:41 crc kubenswrapper[4872]: I1009 08:39:41.279364 4872 generic.go:334] "Generic (PLEG): container finished" podID="1d643d17-4e67-41f2-a8fc-f9790033d364" containerID="4d069edb76e82123a74b53da205732ab4ffc0661bdfd0478962d2718e7c8d49f" exitCode=0 Oct 09 08:39:41 crc kubenswrapper[4872]: I1009 08:39:41.279457 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" event={"ID":"1d643d17-4e67-41f2-a8fc-f9790033d364","Type":"ContainerDied","Data":"4d069edb76e82123a74b53da205732ab4ffc0661bdfd0478962d2718e7c8d49f"} Oct 09 08:39:41 crc kubenswrapper[4872]: I1009 08:39:41.279754 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" event={"ID":"1d643d17-4e67-41f2-a8fc-f9790033d364","Type":"ContainerStarted","Data":"fb0bf155d8f663315e044297efdfd5f7d9454d805c913d69f225598d90571a6f"} Oct 09 08:39:42 crc kubenswrapper[4872]: I1009 08:39:42.297118 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" event={"ID":"1d643d17-4e67-41f2-a8fc-f9790033d364","Type":"ContainerStarted","Data":"1203ed38fed86968a86d44862c31d6e9f066950e0e2348042121eb67907cfdd3"} Oct 09 08:39:42 crc kubenswrapper[4872]: I1009 08:39:42.297292 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:42 crc kubenswrapper[4872]: I1009 08:39:42.346304 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" podStartSLOduration=3.346278003 podStartE2EDuration="3.346278003s" podCreationTimestamp="2025-10-09 08:39:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:39:42.329998911 +0000 UTC m=+1340.520527627" watchObservedRunningTime="2025-10-09 08:39:42.346278003 +0000 UTC m=+1340.536806669" Oct 09 08:39:44 crc kubenswrapper[4872]: I1009 08:39:44.800332 4872 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c7b6c5df9-d4k6h" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.200:5353: i/o timeout" Oct 09 08:39:49 crc kubenswrapper[4872]: I1009 08:39:49.874691 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8c6f6df99-bd65m" Oct 09 08:39:49 crc kubenswrapper[4872]: I1009 08:39:49.928283 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:49 crc kubenswrapper[4872]: I1009 08:39:49.928535 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="dnsmasq-dns" containerID="cri-o://681e731370014a39b6a07c191b12504f3a40e0d0ff39f36f69cd2a0c1fecfc7c" gracePeriod=10 Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.385964 4872 generic.go:334] "Generic (PLEG): container finished" podID="aa161414-76f8-44c4-af75-c2ca575658aa" containerID="681e731370014a39b6a07c191b12504f3a40e0d0ff39f36f69cd2a0c1fecfc7c" exitCode=0 Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.386062 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" event={"ID":"aa161414-76f8-44c4-af75-c2ca575658aa","Type":"ContainerDied","Data":"681e731370014a39b6a07c191b12504f3a40e0d0ff39f36f69cd2a0c1fecfc7c"} Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.386253 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" event={"ID":"aa161414-76f8-44c4-af75-c2ca575658aa","Type":"ContainerDied","Data":"8b1ba23828c5ac6a2d5c658e217a41d26522887e41be5a39cf810f08a7774f28"} Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.386270 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b1ba23828c5ac6a2d5c658e217a41d26522887e41be5a39cf810f08a7774f28" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.457219 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.514761 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515027 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k9jvc\" (UniqueName: \"kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515093 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515157 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515264 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515295 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.515346 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb\") pod \"aa161414-76f8-44c4-af75-c2ca575658aa\" (UID: \"aa161414-76f8-44c4-af75-c2ca575658aa\") " Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.537239 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc" (OuterVolumeSpecName: "kube-api-access-k9jvc") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "kube-api-access-k9jvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.602299 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.606880 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.607495 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.617518 4872 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.617541 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k9jvc\" (UniqueName: \"kubernetes.io/projected/aa161414-76f8-44c4-af75-c2ca575658aa-kube-api-access-k9jvc\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.617552 4872 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.617562 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.622557 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.622966 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.645918 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config" (OuterVolumeSpecName: "config") pod "aa161414-76f8-44c4-af75-c2ca575658aa" (UID: "aa161414-76f8-44c4-af75-c2ca575658aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.719054 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.719095 4872 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:50 crc kubenswrapper[4872]: I1009 08:39:50.719105 4872 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa161414-76f8-44c4-af75-c2ca575658aa-config\") on node \"crc\" DevicePath \"\"" Oct 09 08:39:51 crc kubenswrapper[4872]: I1009 08:39:51.395582 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5576978c7c-msgzg" Oct 09 08:39:51 crc kubenswrapper[4872]: I1009 08:39:51.428347 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:51 crc kubenswrapper[4872]: I1009 08:39:51.436249 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5576978c7c-msgzg"] Oct 09 08:39:52 crc kubenswrapper[4872]: I1009 08:39:52.490622 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" path="/var/lib/kubelet/pods/aa161414-76f8-44c4-af75-c2ca575658aa/volumes" Oct 09 08:40:01 crc kubenswrapper[4872]: I1009 08:40:01.492748 4872 generic.go:334] "Generic (PLEG): container finished" podID="b180311b-aa96-4a46-bbb2-5c32d18b0593" containerID="a02c9dbf41d7104220acf8379aba57674f2817d318c4b709143654a79f7b485b" exitCode=0 Oct 09 08:40:01 crc kubenswrapper[4872]: I1009 08:40:01.492838 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b180311b-aa96-4a46-bbb2-5c32d18b0593","Type":"ContainerDied","Data":"a02c9dbf41d7104220acf8379aba57674f2817d318c4b709143654a79f7b485b"} Oct 09 08:40:02 crc kubenswrapper[4872]: I1009 08:40:02.505182 4872 generic.go:334] "Generic (PLEG): container finished" podID="06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc" containerID="5d5262403109bedd850dfc8654cdb60ad8e8a618174df50bc978500e60a58aca" exitCode=0 Oct 09 08:40:02 crc kubenswrapper[4872]: I1009 08:40:02.505284 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc","Type":"ContainerDied","Data":"5d5262403109bedd850dfc8654cdb60ad8e8a618174df50bc978500e60a58aca"} Oct 09 08:40:02 crc kubenswrapper[4872]: I1009 08:40:02.509093 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b180311b-aa96-4a46-bbb2-5c32d18b0593","Type":"ContainerStarted","Data":"09c2c0434b290396097f2d9b69b3e0d4d3bae2fddb8421c3e7eae86f64b263ab"} Oct 09 08:40:02 crc kubenswrapper[4872]: I1009 08:40:02.509303 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 09 08:40:02 crc kubenswrapper[4872]: I1009 08:40:02.562938 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.562915732 podStartE2EDuration="36.562915732s" podCreationTimestamp="2025-10-09 08:39:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:40:02.558531224 +0000 UTC m=+1360.749059860" watchObservedRunningTime="2025-10-09 08:40:02.562915732 +0000 UTC m=+1360.753444358" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.168397 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6"] Oct 09 08:40:03 crc kubenswrapper[4872]: E1009 08:40:03.169302 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169326 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: E1009 08:40:03.169339 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="init" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169347 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="init" Oct 09 08:40:03 crc kubenswrapper[4872]: E1009 08:40:03.169369 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="init" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169377 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="init" Oct 09 08:40:03 crc kubenswrapper[4872]: E1009 08:40:03.169407 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169414 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169674 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa161414-76f8-44c4-af75-c2ca575658aa" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.169698 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a65d363b-6367-4f29-9d54-9270f6122112" containerName="dnsmasq-dns" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.170453 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.173920 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.174933 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.174997 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.174936 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.188305 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6"] Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.270178 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.270239 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.270258 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.270284 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7q7c\" (UniqueName: \"kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.372288 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.372366 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.372391 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.372424 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7q7c\" (UniqueName: \"kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.377121 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.377392 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.391301 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.395626 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7q7c\" (UniqueName: \"kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.492961 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.520153 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc","Type":"ContainerStarted","Data":"1957a93ae28bbf34b0552a48de7355c781d462bb5979a1d348551ae326d2b368"} Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.520871 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:40:03 crc kubenswrapper[4872]: I1009 08:40:03.549259 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.549241719 podStartE2EDuration="36.549241719s" podCreationTimestamp="2025-10-09 08:39:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 08:40:03.543898194 +0000 UTC m=+1361.734426830" watchObservedRunningTime="2025-10-09 08:40:03.549241719 +0000 UTC m=+1361.739770345" Oct 09 08:40:04 crc kubenswrapper[4872]: I1009 08:40:04.033531 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6"] Oct 09 08:40:04 crc kubenswrapper[4872]: W1009 08:40:04.039492 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb21bf250_3cfe_47da_8ff1_2c4a54dc7b77.slice/crio-8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a WatchSource:0}: Error finding container 8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a: Status 404 returned error can't find the container with id 8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a Oct 09 08:40:04 crc kubenswrapper[4872]: I1009 08:40:04.043411 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:40:04 crc kubenswrapper[4872]: I1009 08:40:04.532037 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" event={"ID":"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77","Type":"ContainerStarted","Data":"8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a"} Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.157371 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.158012 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.158069 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.158974 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.159043 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f" gracePeriod=600 Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.574035 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f" exitCode=0 Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.574108 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f"} Oct 09 08:40:08 crc kubenswrapper[4872]: I1009 08:40:08.574142 4872 scope.go:117] "RemoveContainer" containerID="439bea6e7780c23dd33c763c4fb0d3323318c79552fe3818bbed94847f206451" Oct 09 08:40:13 crc kubenswrapper[4872]: I1009 08:40:13.624569 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549"} Oct 09 08:40:13 crc kubenswrapper[4872]: I1009 08:40:13.627203 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" event={"ID":"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77","Type":"ContainerStarted","Data":"e428af1dfc5c8a5832fef497a3e3ff310262a2e25e45a8f6153b1aa7da8e454f"} Oct 09 08:40:13 crc kubenswrapper[4872]: I1009 08:40:13.664559 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" podStartSLOduration=1.8316766740000001 podStartE2EDuration="10.66454175s" podCreationTimestamp="2025-10-09 08:40:03 +0000 UTC" firstStartedPulling="2025-10-09 08:40:04.04299834 +0000 UTC m=+1362.233526966" lastFinishedPulling="2025-10-09 08:40:12.875863416 +0000 UTC m=+1371.066392042" observedRunningTime="2025-10-09 08:40:13.660503053 +0000 UTC m=+1371.851031679" watchObservedRunningTime="2025-10-09 08:40:13.66454175 +0000 UTC m=+1371.855070386" Oct 09 08:40:16 crc kubenswrapper[4872]: I1009 08:40:16.610597 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 09 08:40:17 crc kubenswrapper[4872]: I1009 08:40:17.627879 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 09 08:40:23 crc kubenswrapper[4872]: I1009 08:40:23.736359 4872 generic.go:334] "Generic (PLEG): container finished" podID="b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" containerID="e428af1dfc5c8a5832fef497a3e3ff310262a2e25e45a8f6153b1aa7da8e454f" exitCode=0 Oct 09 08:40:23 crc kubenswrapper[4872]: I1009 08:40:23.736422 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" event={"ID":"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77","Type":"ContainerDied","Data":"e428af1dfc5c8a5832fef497a3e3ff310262a2e25e45a8f6153b1aa7da8e454f"} Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.158840 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.253170 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle\") pod \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.253328 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory\") pod \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.253431 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key\") pod \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.253512 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7q7c\" (UniqueName: \"kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c\") pod \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\" (UID: \"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77\") " Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.259047 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" (UID: "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.263938 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c" (OuterVolumeSpecName: "kube-api-access-p7q7c") pod "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" (UID: "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77"). InnerVolumeSpecName "kube-api-access-p7q7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.281735 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory" (OuterVolumeSpecName: "inventory") pod "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" (UID: "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.311902 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" (UID: "b21bf250-3cfe-47da-8ff1-2c4a54dc7b77"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.356192 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.356250 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7q7c\" (UniqueName: \"kubernetes.io/projected/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-kube-api-access-p7q7c\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.356273 4872 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.356319 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b21bf250-3cfe-47da-8ff1-2c4a54dc7b77-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.761139 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" event={"ID":"b21bf250-3cfe-47da-8ff1-2c4a54dc7b77","Type":"ContainerDied","Data":"8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a"} Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.761807 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d4fb82246b6c3f470503a2b2d9f20867310c963a044a4913f68dc0d271b0f1a" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.761170 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.839157 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg"] Oct 09 08:40:25 crc kubenswrapper[4872]: E1009 08:40:25.839541 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.839562 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.839787 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21bf250-3cfe-47da-8ff1-2c4a54dc7b77" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.840388 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.842681 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.843195 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.843906 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.843962 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.853373 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg"] Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.972437 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pnrf\" (UniqueName: \"kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.972515 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:25 crc kubenswrapper[4872]: I1009 08:40:25.972566 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.074525 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.074604 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.074740 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pnrf\" (UniqueName: \"kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.079617 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.081688 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.095917 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pnrf\" (UniqueName: \"kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-qd5jg\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.168096 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.501681 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg"] Oct 09 08:40:26 crc kubenswrapper[4872]: W1009 08:40:26.503343 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bf8bedf_9b23_4570_b728_fc01c39c9373.slice/crio-d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e WatchSource:0}: Error finding container d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e: Status 404 returned error can't find the container with id d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e Oct 09 08:40:26 crc kubenswrapper[4872]: I1009 08:40:26.776906 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" event={"ID":"5bf8bedf-9b23-4570-b728-fc01c39c9373","Type":"ContainerStarted","Data":"d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e"} Oct 09 08:40:27 crc kubenswrapper[4872]: I1009 08:40:27.790626 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" event={"ID":"5bf8bedf-9b23-4570-b728-fc01c39c9373","Type":"ContainerStarted","Data":"1358e0cb7ded82ee4a7cc015ce02e8890cebb8a98d585d791279a66608749a27"} Oct 09 08:40:27 crc kubenswrapper[4872]: I1009 08:40:27.811816 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" podStartSLOduration=2.045736065 podStartE2EDuration="2.811798074s" podCreationTimestamp="2025-10-09 08:40:25 +0000 UTC" firstStartedPulling="2025-10-09 08:40:26.505152756 +0000 UTC m=+1384.695681382" lastFinishedPulling="2025-10-09 08:40:27.271214765 +0000 UTC m=+1385.461743391" observedRunningTime="2025-10-09 08:40:27.807320364 +0000 UTC m=+1385.997849010" watchObservedRunningTime="2025-10-09 08:40:27.811798074 +0000 UTC m=+1386.002326700" Oct 09 08:40:30 crc kubenswrapper[4872]: I1009 08:40:30.828089 4872 generic.go:334] "Generic (PLEG): container finished" podID="5bf8bedf-9b23-4570-b728-fc01c39c9373" containerID="1358e0cb7ded82ee4a7cc015ce02e8890cebb8a98d585d791279a66608749a27" exitCode=0 Oct 09 08:40:30 crc kubenswrapper[4872]: I1009 08:40:30.828198 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" event={"ID":"5bf8bedf-9b23-4570-b728-fc01c39c9373","Type":"ContainerDied","Data":"1358e0cb7ded82ee4a7cc015ce02e8890cebb8a98d585d791279a66608749a27"} Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.208986 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.298991 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key\") pod \"5bf8bedf-9b23-4570-b728-fc01c39c9373\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.299088 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory\") pod \"5bf8bedf-9b23-4570-b728-fc01c39c9373\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.299113 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pnrf\" (UniqueName: \"kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf\") pod \"5bf8bedf-9b23-4570-b728-fc01c39c9373\" (UID: \"5bf8bedf-9b23-4570-b728-fc01c39c9373\") " Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.305879 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf" (OuterVolumeSpecName: "kube-api-access-2pnrf") pod "5bf8bedf-9b23-4570-b728-fc01c39c9373" (UID: "5bf8bedf-9b23-4570-b728-fc01c39c9373"). InnerVolumeSpecName "kube-api-access-2pnrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.335333 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5bf8bedf-9b23-4570-b728-fc01c39c9373" (UID: "5bf8bedf-9b23-4570-b728-fc01c39c9373"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.367661 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory" (OuterVolumeSpecName: "inventory") pod "5bf8bedf-9b23-4570-b728-fc01c39c9373" (UID: "5bf8bedf-9b23-4570-b728-fc01c39c9373"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.402277 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.402347 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5bf8bedf-9b23-4570-b728-fc01c39c9373-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.402367 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pnrf\" (UniqueName: \"kubernetes.io/projected/5bf8bedf-9b23-4570-b728-fc01c39c9373-kube-api-access-2pnrf\") on node \"crc\" DevicePath \"\"" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.852380 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" event={"ID":"5bf8bedf-9b23-4570-b728-fc01c39c9373","Type":"ContainerDied","Data":"d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e"} Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.852892 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8ba13f4a254f18fa481f864993254ae9f993ca4d0679465d7e47a3ab253485e" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.852526 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-qd5jg" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.931917 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx"] Oct 09 08:40:32 crc kubenswrapper[4872]: E1009 08:40:32.932624 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5bf8bedf-9b23-4570-b728-fc01c39c9373" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.932778 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="5bf8bedf-9b23-4570-b728-fc01c39c9373" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.933046 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="5bf8bedf-9b23-4570-b728-fc01c39c9373" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.933891 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.940095 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.940347 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.940419 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.940763 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:40:32 crc kubenswrapper[4872]: I1009 08:40:32.949888 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx"] Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.012519 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.012579 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.012751 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.012795 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2bbh\" (UniqueName: \"kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.115013 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2bbh\" (UniqueName: \"kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.115111 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.115151 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.115364 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.120368 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.121491 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.125996 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.131810 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2bbh\" (UniqueName: \"kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.250627 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:40:33 crc kubenswrapper[4872]: W1009 08:40:33.786023 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod441768e0_9ee7_4d91_a9e1_d72ef8fb47e3.slice/crio-25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf WatchSource:0}: Error finding container 25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf: Status 404 returned error can't find the container with id 25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.790129 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx"] Oct 09 08:40:33 crc kubenswrapper[4872]: I1009 08:40:33.867915 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" event={"ID":"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3","Type":"ContainerStarted","Data":"25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf"} Oct 09 08:40:34 crc kubenswrapper[4872]: I1009 08:40:34.880102 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" event={"ID":"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3","Type":"ContainerStarted","Data":"a8bd704e1a3b45893f8792f7b2393fe0e4bca6b8eb6e649879ff25d9996235d2"} Oct 09 08:40:34 crc kubenswrapper[4872]: I1009 08:40:34.907252 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" podStartSLOduration=2.379568774 podStartE2EDuration="2.907179146s" podCreationTimestamp="2025-10-09 08:40:32 +0000 UTC" firstStartedPulling="2025-10-09 08:40:33.789770987 +0000 UTC m=+1391.980299623" lastFinishedPulling="2025-10-09 08:40:34.317381379 +0000 UTC m=+1392.507909995" observedRunningTime="2025-10-09 08:40:34.900848092 +0000 UTC m=+1393.091376728" watchObservedRunningTime="2025-10-09 08:40:34.907179146 +0000 UTC m=+1393.097707782" Oct 09 08:40:45 crc kubenswrapper[4872]: I1009 08:40:45.169200 4872 scope.go:117] "RemoveContainer" containerID="e8fa4656d598511564adf0e133ae4d8bb506b76e5cfa04100134020a2b209312" Oct 09 08:40:45 crc kubenswrapper[4872]: I1009 08:40:45.200127 4872 scope.go:117] "RemoveContainer" containerID="4d374843dc49fd7757ddcf08ee23c47d6bb3f10cabe4e1deb7a461f2b2087d12" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.286203 4872 scope.go:117] "RemoveContainer" containerID="5c83c3b24f32f5e798e35bf466c8d371ec7a7489108dacfe9944717d98ef6cde" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.314991 4872 scope.go:117] "RemoveContainer" containerID="116426969a0cf57d52722140e2e2d90d9cb4d2a6fc17db08e75e63c689a96d8e" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.373536 4872 scope.go:117] "RemoveContainer" containerID="dce5e5b0b0f2782d02c18f324b8c3a56ae44440a849738813925852f7f3d37eb" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.396759 4872 scope.go:117] "RemoveContainer" containerID="401c2b173285cfab1735af4da1a043607f76c7e975f58eb132c70a3a0b12beaf" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.415284 4872 scope.go:117] "RemoveContainer" containerID="2744602f9826eba65667e9ba0edc1788b8a741e19c0e765e42b227cbede4f2e9" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.449762 4872 scope.go:117] "RemoveContainer" containerID="15af2b37f0c78a931959ac54657df5857e1e3fdd70618ebb77e7d2e59d5a0243" Oct 09 08:41:45 crc kubenswrapper[4872]: I1009 08:41:45.483670 4872 scope.go:117] "RemoveContainer" containerID="fa7e50b7bf9d20e414407cefeec9c0a3953becbf8defe0943a0370c0802cdafb" Oct 09 08:41:47 crc kubenswrapper[4872]: I1009 08:41:47.922553 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:41:47 crc kubenswrapper[4872]: I1009 08:41:47.924976 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:47 crc kubenswrapper[4872]: I1009 08:41:47.935150 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.025762 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.025835 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvtpj\" (UniqueName: \"kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.025941 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.128035 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvtpj\" (UniqueName: \"kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.128097 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.128226 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.128703 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.128723 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.162839 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvtpj\" (UniqueName: \"kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj\") pod \"certified-operators-7rnwx\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.253024 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:48 crc kubenswrapper[4872]: I1009 08:41:48.773106 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:41:49 crc kubenswrapper[4872]: I1009 08:41:49.625873 4872 generic.go:334] "Generic (PLEG): container finished" podID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerID="ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b" exitCode=0 Oct 09 08:41:49 crc kubenswrapper[4872]: I1009 08:41:49.625932 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerDied","Data":"ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b"} Oct 09 08:41:49 crc kubenswrapper[4872]: I1009 08:41:49.626219 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerStarted","Data":"341b3ee3aecb5fed0d64e7fcc5c35e24fce81a19c778b5dc1638773b5c0eea18"} Oct 09 08:41:51 crc kubenswrapper[4872]: I1009 08:41:51.648266 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerStarted","Data":"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5"} Oct 09 08:41:52 crc kubenswrapper[4872]: I1009 08:41:52.662956 4872 generic.go:334] "Generic (PLEG): container finished" podID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerID="c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5" exitCode=0 Oct 09 08:41:52 crc kubenswrapper[4872]: I1009 08:41:52.663014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerDied","Data":"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5"} Oct 09 08:41:54 crc kubenswrapper[4872]: I1009 08:41:54.694126 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerStarted","Data":"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86"} Oct 09 08:41:54 crc kubenswrapper[4872]: I1009 08:41:54.720210 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7rnwx" podStartSLOduration=3.828481391 podStartE2EDuration="7.720190011s" podCreationTimestamp="2025-10-09 08:41:47 +0000 UTC" firstStartedPulling="2025-10-09 08:41:49.629503588 +0000 UTC m=+1467.820032224" lastFinishedPulling="2025-10-09 08:41:53.521212178 +0000 UTC m=+1471.711740844" observedRunningTime="2025-10-09 08:41:54.712067584 +0000 UTC m=+1472.902596210" watchObservedRunningTime="2025-10-09 08:41:54.720190011 +0000 UTC m=+1472.910718637" Oct 09 08:41:58 crc kubenswrapper[4872]: I1009 08:41:58.253344 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:58 crc kubenswrapper[4872]: I1009 08:41:58.253882 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:58 crc kubenswrapper[4872]: I1009 08:41:58.320401 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:58 crc kubenswrapper[4872]: I1009 08:41:58.793235 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:41:58 crc kubenswrapper[4872]: I1009 08:41:58.848189 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:42:00 crc kubenswrapper[4872]: I1009 08:42:00.745690 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7rnwx" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="registry-server" containerID="cri-o://2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86" gracePeriod=2 Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.182973 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.283926 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities\") pod \"e0ff1b11-d25f-4d06-8873-486c4c58f090\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.284263 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content\") pod \"e0ff1b11-d25f-4d06-8873-486c4c58f090\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.284301 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvtpj\" (UniqueName: \"kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj\") pod \"e0ff1b11-d25f-4d06-8873-486c4c58f090\" (UID: \"e0ff1b11-d25f-4d06-8873-486c4c58f090\") " Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.285243 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities" (OuterVolumeSpecName: "utilities") pod "e0ff1b11-d25f-4d06-8873-486c4c58f090" (UID: "e0ff1b11-d25f-4d06-8873-486c4c58f090"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.299122 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj" (OuterVolumeSpecName: "kube-api-access-rvtpj") pod "e0ff1b11-d25f-4d06-8873-486c4c58f090" (UID: "e0ff1b11-d25f-4d06-8873-486c4c58f090"). InnerVolumeSpecName "kube-api-access-rvtpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.339857 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0ff1b11-d25f-4d06-8873-486c4c58f090" (UID: "e0ff1b11-d25f-4d06-8873-486c4c58f090"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.386180 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvtpj\" (UniqueName: \"kubernetes.io/projected/e0ff1b11-d25f-4d06-8873-486c4c58f090-kube-api-access-rvtpj\") on node \"crc\" DevicePath \"\"" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.386219 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.386231 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0ff1b11-d25f-4d06-8873-486c4c58f090-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.759351 4872 generic.go:334] "Generic (PLEG): container finished" podID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerID="2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86" exitCode=0 Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.759458 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerDied","Data":"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86"} Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.759693 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rnwx" event={"ID":"e0ff1b11-d25f-4d06-8873-486c4c58f090","Type":"ContainerDied","Data":"341b3ee3aecb5fed0d64e7fcc5c35e24fce81a19c778b5dc1638773b5c0eea18"} Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.759714 4872 scope.go:117] "RemoveContainer" containerID="2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.759492 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rnwx" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.789917 4872 scope.go:117] "RemoveContainer" containerID="c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.806053 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.815486 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7rnwx"] Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.833671 4872 scope.go:117] "RemoveContainer" containerID="ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.867246 4872 scope.go:117] "RemoveContainer" containerID="2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86" Oct 09 08:42:01 crc kubenswrapper[4872]: E1009 08:42:01.867708 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86\": container with ID starting with 2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86 not found: ID does not exist" containerID="2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.867747 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86"} err="failed to get container status \"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86\": rpc error: code = NotFound desc = could not find container \"2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86\": container with ID starting with 2dfe742984efc9e46c23a3553d618d31cd8bfd1b05bc8bda20ee2d961c2d5c86 not found: ID does not exist" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.867791 4872 scope.go:117] "RemoveContainer" containerID="c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5" Oct 09 08:42:01 crc kubenswrapper[4872]: E1009 08:42:01.868246 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5\": container with ID starting with c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5 not found: ID does not exist" containerID="c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.868297 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5"} err="failed to get container status \"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5\": rpc error: code = NotFound desc = could not find container \"c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5\": container with ID starting with c0aab86a74c8ed37d3e7f10a46f9016ba541037cbb9a84f0d54ef8e29138a4c5 not found: ID does not exist" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.868315 4872 scope.go:117] "RemoveContainer" containerID="ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b" Oct 09 08:42:01 crc kubenswrapper[4872]: E1009 08:42:01.868579 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b\": container with ID starting with ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b not found: ID does not exist" containerID="ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b" Oct 09 08:42:01 crc kubenswrapper[4872]: I1009 08:42:01.868605 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b"} err="failed to get container status \"ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b\": rpc error: code = NotFound desc = could not find container \"ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b\": container with ID starting with ad6e22d0bbd8c82e9a559dfaf2eb7adf1c650d225a02e260eab000ac7419988b not found: ID does not exist" Oct 09 08:42:02 crc kubenswrapper[4872]: I1009 08:42:02.475759 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" path="/var/lib/kubelet/pods/e0ff1b11-d25f-4d06-8873-486c4c58f090/volumes" Oct 09 08:42:38 crc kubenswrapper[4872]: I1009 08:42:38.156660 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:42:38 crc kubenswrapper[4872]: I1009 08:42:38.157295 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:42:45 crc kubenswrapper[4872]: I1009 08:42:45.749705 4872 scope.go:117] "RemoveContainer" containerID="f0a4580a74aac6aeca707abe71a6c3e3412ca12ba56340733515caa6cf65c087" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.457098 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:01 crc kubenswrapper[4872]: E1009 08:43:01.458071 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="registry-server" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.458088 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="registry-server" Oct 09 08:43:01 crc kubenswrapper[4872]: E1009 08:43:01.458114 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="extract-utilities" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.458123 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="extract-utilities" Oct 09 08:43:01 crc kubenswrapper[4872]: E1009 08:43:01.458138 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="extract-content" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.458146 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="extract-content" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.458374 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0ff1b11-d25f-4d06-8873-486c4c58f090" containerName="registry-server" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.460876 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.481328 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.586170 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.586414 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.586520 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsm5j\" (UniqueName: \"kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.688552 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.688737 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.688796 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsm5j\" (UniqueName: \"kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.689308 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.689368 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.706503 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsm5j\" (UniqueName: \"kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j\") pod \"community-operators-bh826\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:01 crc kubenswrapper[4872]: I1009 08:43:01.786984 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:02 crc kubenswrapper[4872]: W1009 08:43:02.255104 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37ae0c1a_a5ed_44d5_b777_05f5acd24031.slice/crio-259a79511c16116ae1012c40438d4aee159e0a119df4c39326d79730c328995d WatchSource:0}: Error finding container 259a79511c16116ae1012c40438d4aee159e0a119df4c39326d79730c328995d: Status 404 returned error can't find the container with id 259a79511c16116ae1012c40438d4aee159e0a119df4c39326d79730c328995d Oct 09 08:43:02 crc kubenswrapper[4872]: I1009 08:43:02.256586 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:02 crc kubenswrapper[4872]: I1009 08:43:02.350385 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerStarted","Data":"259a79511c16116ae1012c40438d4aee159e0a119df4c39326d79730c328995d"} Oct 09 08:43:03 crc kubenswrapper[4872]: I1009 08:43:03.361896 4872 generic.go:334] "Generic (PLEG): container finished" podID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerID="26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2" exitCode=0 Oct 09 08:43:03 crc kubenswrapper[4872]: I1009 08:43:03.362001 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerDied","Data":"26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2"} Oct 09 08:43:04 crc kubenswrapper[4872]: I1009 08:43:04.374693 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerStarted","Data":"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101"} Oct 09 08:43:05 crc kubenswrapper[4872]: I1009 08:43:05.390450 4872 generic.go:334] "Generic (PLEG): container finished" podID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerID="b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101" exitCode=0 Oct 09 08:43:05 crc kubenswrapper[4872]: I1009 08:43:05.390537 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerDied","Data":"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101"} Oct 09 08:43:06 crc kubenswrapper[4872]: I1009 08:43:06.404224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerStarted","Data":"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d"} Oct 09 08:43:06 crc kubenswrapper[4872]: I1009 08:43:06.429979 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bh826" podStartSLOduration=2.6502942860000003 podStartE2EDuration="5.429955195s" podCreationTimestamp="2025-10-09 08:43:01 +0000 UTC" firstStartedPulling="2025-10-09 08:43:03.365896164 +0000 UTC m=+1541.556424840" lastFinishedPulling="2025-10-09 08:43:06.145557123 +0000 UTC m=+1544.336085749" observedRunningTime="2025-10-09 08:43:06.425111994 +0000 UTC m=+1544.615640630" watchObservedRunningTime="2025-10-09 08:43:06.429955195 +0000 UTC m=+1544.620483821" Oct 09 08:43:08 crc kubenswrapper[4872]: I1009 08:43:08.156723 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:43:08 crc kubenswrapper[4872]: I1009 08:43:08.156795 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:43:11 crc kubenswrapper[4872]: I1009 08:43:11.787773 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:11 crc kubenswrapper[4872]: I1009 08:43:11.788387 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:11 crc kubenswrapper[4872]: I1009 08:43:11.832009 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:12 crc kubenswrapper[4872]: I1009 08:43:12.518947 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:12 crc kubenswrapper[4872]: I1009 08:43:12.569221 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.480076 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bh826" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="registry-server" containerID="cri-o://a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d" gracePeriod=2 Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.935138 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.970174 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities\") pod \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.970397 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsm5j\" (UniqueName: \"kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j\") pod \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.970747 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content\") pod \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\" (UID: \"37ae0c1a-a5ed-44d5-b777-05f5acd24031\") " Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.971740 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities" (OuterVolumeSpecName: "utilities") pod "37ae0c1a-a5ed-44d5-b777-05f5acd24031" (UID: "37ae0c1a-a5ed-44d5-b777-05f5acd24031"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:43:14 crc kubenswrapper[4872]: I1009 08:43:14.977295 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j" (OuterVolumeSpecName: "kube-api-access-rsm5j") pod "37ae0c1a-a5ed-44d5-b777-05f5acd24031" (UID: "37ae0c1a-a5ed-44d5-b777-05f5acd24031"). InnerVolumeSpecName "kube-api-access-rsm5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.033968 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "37ae0c1a-a5ed-44d5-b777-05f5acd24031" (UID: "37ae0c1a-a5ed-44d5-b777-05f5acd24031"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.073189 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.073223 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsm5j\" (UniqueName: \"kubernetes.io/projected/37ae0c1a-a5ed-44d5-b777-05f5acd24031-kube-api-access-rsm5j\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.073234 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/37ae0c1a-a5ed-44d5-b777-05f5acd24031-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.492735 4872 generic.go:334] "Generic (PLEG): container finished" podID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerID="a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d" exitCode=0 Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.492790 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerDied","Data":"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d"} Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.492815 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bh826" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.492828 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bh826" event={"ID":"37ae0c1a-a5ed-44d5-b777-05f5acd24031","Type":"ContainerDied","Data":"259a79511c16116ae1012c40438d4aee159e0a119df4c39326d79730c328995d"} Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.492854 4872 scope.go:117] "RemoveContainer" containerID="a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.530663 4872 scope.go:117] "RemoveContainer" containerID="b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.536343 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.546604 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bh826"] Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.558420 4872 scope.go:117] "RemoveContainer" containerID="26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.609424 4872 scope.go:117] "RemoveContainer" containerID="a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d" Oct 09 08:43:15 crc kubenswrapper[4872]: E1009 08:43:15.609902 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d\": container with ID starting with a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d not found: ID does not exist" containerID="a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.609938 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d"} err="failed to get container status \"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d\": rpc error: code = NotFound desc = could not find container \"a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d\": container with ID starting with a86eeda38549e8be0ce421e6d8a4e148f45f0c2118909e25b1fab652ba42287d not found: ID does not exist" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.609964 4872 scope.go:117] "RemoveContainer" containerID="b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101" Oct 09 08:43:15 crc kubenswrapper[4872]: E1009 08:43:15.611008 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101\": container with ID starting with b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101 not found: ID does not exist" containerID="b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.611076 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101"} err="failed to get container status \"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101\": rpc error: code = NotFound desc = could not find container \"b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101\": container with ID starting with b84539bc9dcfe64aba3609398c4b4352a05d2f5e68f43aea3d8c62f8ee366101 not found: ID does not exist" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.611117 4872 scope.go:117] "RemoveContainer" containerID="26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2" Oct 09 08:43:15 crc kubenswrapper[4872]: E1009 08:43:15.611588 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2\": container with ID starting with 26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2 not found: ID does not exist" containerID="26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2" Oct 09 08:43:15 crc kubenswrapper[4872]: I1009 08:43:15.611693 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2"} err="failed to get container status \"26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2\": rpc error: code = NotFound desc = could not find container \"26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2\": container with ID starting with 26b8e1a6d803324c2fde2d422dc5789adb8fbe63d5bdc68f8e9e6e97fcb38ae2 not found: ID does not exist" Oct 09 08:43:16 crc kubenswrapper[4872]: I1009 08:43:16.485306 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" path="/var/lib/kubelet/pods/37ae0c1a-a5ed-44d5-b777-05f5acd24031/volumes" Oct 09 08:43:33 crc kubenswrapper[4872]: I1009 08:43:33.680035 4872 generic.go:334] "Generic (PLEG): container finished" podID="441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" containerID="a8bd704e1a3b45893f8792f7b2393fe0e4bca6b8eb6e649879ff25d9996235d2" exitCode=0 Oct 09 08:43:33 crc kubenswrapper[4872]: I1009 08:43:33.680115 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" event={"ID":"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3","Type":"ContainerDied","Data":"a8bd704e1a3b45893f8792f7b2393fe0e4bca6b8eb6e649879ff25d9996235d2"} Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.117796 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.267991 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory\") pod \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.268098 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key\") pod \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.268149 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle\") pod \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.268181 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2bbh\" (UniqueName: \"kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh\") pod \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\" (UID: \"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3\") " Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.274673 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" (UID: "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.277086 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh" (OuterVolumeSpecName: "kube-api-access-q2bbh") pod "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" (UID: "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3"). InnerVolumeSpecName "kube-api-access-q2bbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.300909 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" (UID: "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.301268 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory" (OuterVolumeSpecName: "inventory") pod "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" (UID: "441768e0-9ee7-4d91-a9e1-d72ef8fb47e3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.370192 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.370243 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.370257 4872 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.370273 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2bbh\" (UniqueName: \"kubernetes.io/projected/441768e0-9ee7-4d91-a9e1-d72ef8fb47e3-kube-api-access-q2bbh\") on node \"crc\" DevicePath \"\"" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.708974 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" event={"ID":"441768e0-9ee7-4d91-a9e1-d72ef8fb47e3","Type":"ContainerDied","Data":"25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf"} Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.709317 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25e48b3d152e84e67fba341cf7fa57a3a098f497bb52599664bcf372e3d7fadf" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.709089 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.794738 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt"] Oct 09 08:43:35 crc kubenswrapper[4872]: E1009 08:43:35.795151 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="extract-content" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.795172 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="extract-content" Oct 09 08:43:35 crc kubenswrapper[4872]: E1009 08:43:35.795185 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.795194 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 08:43:35 crc kubenswrapper[4872]: E1009 08:43:35.795206 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="extract-utilities" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.795212 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="extract-utilities" Oct 09 08:43:35 crc kubenswrapper[4872]: E1009 08:43:35.795239 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="registry-server" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.795246 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="registry-server" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.802880 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="441768e0-9ee7-4d91-a9e1-d72ef8fb47e3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.802968 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ae0c1a-a5ed-44d5-b777-05f5acd24031" containerName="registry-server" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.804216 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.808713 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.813482 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.813804 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.813964 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.828517 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt"] Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.880540 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.881000 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.881174 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s2jd2\" (UniqueName: \"kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.983959 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.984075 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2jd2\" (UniqueName: \"kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.984214 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.992610 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:35 crc kubenswrapper[4872]: I1009 08:43:35.993255 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:36 crc kubenswrapper[4872]: I1009 08:43:36.010661 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2jd2\" (UniqueName: \"kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:36 crc kubenswrapper[4872]: I1009 08:43:36.167302 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:43:36 crc kubenswrapper[4872]: I1009 08:43:36.690495 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt"] Oct 09 08:43:36 crc kubenswrapper[4872]: I1009 08:43:36.724421 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" event={"ID":"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08","Type":"ContainerStarted","Data":"b3221a32e8de4f69b7d1554f6b78ec26e263bdb15b9265cad41478b9f71ad4d7"} Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.157039 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.157608 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.157675 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.158423 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.158479 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" gracePeriod=600 Oct 09 08:43:38 crc kubenswrapper[4872]: E1009 08:43:38.291267 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.746166 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" exitCode=0 Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.746229 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549"} Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.746291 4872 scope.go:117] "RemoveContainer" containerID="6c2589847945b7438460106d4d3dc11a14a1d9b0881d8bc1fcd0954436bbf11f" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.747404 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:43:38 crc kubenswrapper[4872]: E1009 08:43:38.747923 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:43:38 crc kubenswrapper[4872]: I1009 08:43:38.751054 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" event={"ID":"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08","Type":"ContainerStarted","Data":"e2f615a34f8c17349a4ebb82a435e035c2db8d4c55f22d126ec053b510c6d230"} Oct 09 08:43:45 crc kubenswrapper[4872]: I1009 08:43:45.875344 4872 scope.go:117] "RemoveContainer" containerID="e5456a0bc88528481d0e92e61aeb1fa5c43395c2ec951ebab383d1b09cac6bf3" Oct 09 08:43:53 crc kubenswrapper[4872]: I1009 08:43:53.461630 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:43:53 crc kubenswrapper[4872]: E1009 08:43:53.462503 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:06 crc kubenswrapper[4872]: I1009 08:44:06.462051 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:44:06 crc kubenswrapper[4872]: E1009 08:44:06.463117 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:18 crc kubenswrapper[4872]: I1009 08:44:18.462931 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:44:18 crc kubenswrapper[4872]: E1009 08:44:18.464268 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:30 crc kubenswrapper[4872]: I1009 08:44:30.462158 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:44:30 crc kubenswrapper[4872]: E1009 08:44:30.462846 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:33 crc kubenswrapper[4872]: I1009 08:44:33.037176 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" podStartSLOduration=57.106056578 podStartE2EDuration="58.037039807s" podCreationTimestamp="2025-10-09 08:43:35 +0000 UTC" firstStartedPulling="2025-10-09 08:43:36.70843892 +0000 UTC m=+1574.898967536" lastFinishedPulling="2025-10-09 08:43:37.639422139 +0000 UTC m=+1575.829950765" observedRunningTime="2025-10-09 08:43:38.791727672 +0000 UTC m=+1576.982256378" watchObservedRunningTime="2025-10-09 08:44:33.037039807 +0000 UTC m=+1631.227568433" Oct 09 08:44:33 crc kubenswrapper[4872]: I1009 08:44:33.048819 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nxr9k"] Oct 09 08:44:33 crc kubenswrapper[4872]: I1009 08:44:33.058176 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nxr9k"] Oct 09 08:44:34 crc kubenswrapper[4872]: I1009 08:44:34.477361 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e189a93b-9f2f-4c99-b315-69473291ed9c" path="/var/lib/kubelet/pods/e189a93b-9f2f-4c99-b315-69473291ed9c/volumes" Oct 09 08:44:38 crc kubenswrapper[4872]: I1009 08:44:38.043619 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bvqd5"] Oct 09 08:44:38 crc kubenswrapper[4872]: I1009 08:44:38.053933 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bvqd5"] Oct 09 08:44:38 crc kubenswrapper[4872]: I1009 08:44:38.474704 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d67046a4-17d6-489f-857f-f8b016906baa" path="/var/lib/kubelet/pods/d67046a4-17d6-489f-857f-f8b016906baa/volumes" Oct 09 08:44:42 crc kubenswrapper[4872]: I1009 08:44:42.051341 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-98xd2"] Oct 09 08:44:42 crc kubenswrapper[4872]: I1009 08:44:42.064449 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-98xd2"] Oct 09 08:44:42 crc kubenswrapper[4872]: I1009 08:44:42.467770 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:44:42 crc kubenswrapper[4872]: E1009 08:44:42.468323 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:42 crc kubenswrapper[4872]: I1009 08:44:42.475410 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="291de9b6-d184-4fdd-817a-ac679afcfe63" path="/var/lib/kubelet/pods/291de9b6-d184-4fdd-817a-ac679afcfe63/volumes" Oct 09 08:44:43 crc kubenswrapper[4872]: I1009 08:44:43.039675 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-843f-account-create-kpbb9"] Oct 09 08:44:43 crc kubenswrapper[4872]: I1009 08:44:43.051040 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-843f-account-create-kpbb9"] Oct 09 08:44:44 crc kubenswrapper[4872]: I1009 08:44:44.473711 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ed1fcc1-b162-4818-abc1-4674bdf739c3" path="/var/lib/kubelet/pods/1ed1fcc1-b162-4818-abc1-4674bdf739c3/volumes" Oct 09 08:44:45 crc kubenswrapper[4872]: I1009 08:44:45.951966 4872 scope.go:117] "RemoveContainer" containerID="556f0bc9baa9fe0435b1ccc929651c534bdbf7121e229eecc4eb288cc04af131" Oct 09 08:44:45 crc kubenswrapper[4872]: I1009 08:44:45.993941 4872 scope.go:117] "RemoveContainer" containerID="33ee4fbb0d3c8e72f58c573536507b0c73b88f5a15ba3362f453b68cb6fc9a3b" Oct 09 08:44:46 crc kubenswrapper[4872]: I1009 08:44:46.045052 4872 scope.go:117] "RemoveContainer" containerID="f2a6ee52f21536fb1dc6a270480ccc5c38f6d1a4344d77078d462f26453869ae" Oct 09 08:44:46 crc kubenswrapper[4872]: I1009 08:44:46.100209 4872 scope.go:117] "RemoveContainer" containerID="0b3af886e27dd94b6f2fcc82a81b968c48b231534811d8dfbb66ebf33c02e406" Oct 09 08:44:49 crc kubenswrapper[4872]: I1009 08:44:49.036415 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-04fe-account-create-5nhvz"] Oct 09 08:44:49 crc kubenswrapper[4872]: I1009 08:44:49.047456 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-04fe-account-create-5nhvz"] Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.481348 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1" path="/var/lib/kubelet/pods/69acc3cc-d3f1-4ce3-b5ed-f7b5856d0de1/volumes" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.752198 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.754699 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.771943 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.803216 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m59qr\" (UniqueName: \"kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.803301 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.803330 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.906207 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m59qr\" (UniqueName: \"kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.906276 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.906294 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.906917 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.907076 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:50 crc kubenswrapper[4872]: I1009 08:44:50.938954 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m59qr\" (UniqueName: \"kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr\") pod \"redhat-marketplace-ckl88\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:51 crc kubenswrapper[4872]: I1009 08:44:51.093579 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:44:51 crc kubenswrapper[4872]: I1009 08:44:51.573057 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.515031 4872 generic.go:334] "Generic (PLEG): container finished" podID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerID="61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6" exitCode=0 Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.515505 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerDied","Data":"61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6"} Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.516311 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerStarted","Data":"26946e53de4b2f00dd9bbbeb78ce449efe4044996a065a5f1150611be36801b1"} Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.570708 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.574206 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.576252 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.643607 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.643959 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.644160 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptzcq\" (UniqueName: \"kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.746321 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.746425 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.746468 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptzcq\" (UniqueName: \"kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.747151 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.747221 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.767024 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptzcq\" (UniqueName: \"kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq\") pod \"redhat-operators-k2jcv\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:52 crc kubenswrapper[4872]: I1009 08:44:52.900140 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.033079 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2695-account-create-rcqsk"] Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.047700 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2695-account-create-rcqsk"] Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.365426 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.532317 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerStarted","Data":"ff10600965f4d52420b7475a8a1ac528c9984a55a8829c99df9095daa48d3a6e"} Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.537113 4872 generic.go:334] "Generic (PLEG): container finished" podID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerID="db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f" exitCode=0 Oct 09 08:44:53 crc kubenswrapper[4872]: I1009 08:44:53.537179 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerDied","Data":"db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f"} Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.461682 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:44:54 crc kubenswrapper[4872]: E1009 08:44:54.462380 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.471908 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d4694ce-cbc2-4a3a-b46b-08049d1822d9" path="/var/lib/kubelet/pods/0d4694ce-cbc2-4a3a-b46b-08049d1822d9/volumes" Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.557432 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerID="5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a" exitCode=0 Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.557481 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerDied","Data":"5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a"} Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.564902 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerStarted","Data":"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730"} Oct 09 08:44:54 crc kubenswrapper[4872]: I1009 08:44:54.601748 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ckl88" podStartSLOduration=3.117389307 podStartE2EDuration="4.601729889s" podCreationTimestamp="2025-10-09 08:44:50 +0000 UTC" firstStartedPulling="2025-10-09 08:44:52.520152047 +0000 UTC m=+1650.710680713" lastFinishedPulling="2025-10-09 08:44:54.004492669 +0000 UTC m=+1652.195021295" observedRunningTime="2025-10-09 08:44:54.595745374 +0000 UTC m=+1652.786274020" watchObservedRunningTime="2025-10-09 08:44:54.601729889 +0000 UTC m=+1652.792258515" Oct 09 08:44:56 crc kubenswrapper[4872]: I1009 08:44:56.585022 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerID="51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88" exitCode=0 Oct 09 08:44:56 crc kubenswrapper[4872]: I1009 08:44:56.585087 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerDied","Data":"51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88"} Oct 09 08:44:57 crc kubenswrapper[4872]: I1009 08:44:57.599815 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerStarted","Data":"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458"} Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.024918 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k2jcv" podStartSLOduration=5.567069034 podStartE2EDuration="8.024894243s" podCreationTimestamp="2025-10-09 08:44:52 +0000 UTC" firstStartedPulling="2025-10-09 08:44:54.559604552 +0000 UTC m=+1652.750133198" lastFinishedPulling="2025-10-09 08:44:57.017429781 +0000 UTC m=+1655.207958407" observedRunningTime="2025-10-09 08:44:57.618124402 +0000 UTC m=+1655.808653028" watchObservedRunningTime="2025-10-09 08:45:00.024894243 +0000 UTC m=+1658.215422869" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.030783 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-4x54s"] Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.040572 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-4x54s"] Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.153014 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg"] Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.156860 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.163181 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.163255 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.189242 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.189289 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwpdq\" (UniqueName: \"kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.189564 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.190808 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg"] Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.291461 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.291528 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.291561 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwpdq\" (UniqueName: \"kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.293356 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.298534 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.311550 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwpdq\" (UniqueName: \"kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq\") pod \"collect-profiles-29333325-9zvjg\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.476200 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca98354-3506-4c2a-99d3-f482190c4326" path="/var/lib/kubelet/pods/fca98354-3506-4c2a-99d3-f482190c4326/volumes" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.493201 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:00 crc kubenswrapper[4872]: I1009 08:45:00.972586 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg"] Oct 09 08:45:00 crc kubenswrapper[4872]: W1009 08:45:00.976066 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31e6220d_a6e0_4688_9827_7f889fd01fe8.slice/crio-c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec WatchSource:0}: Error finding container c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec: Status 404 returned error can't find the container with id c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.037705 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-8xw9z"] Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.048152 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-cp4rx"] Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.058049 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-cp4rx"] Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.067697 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-8xw9z"] Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.096219 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.096276 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.163434 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.648695 4872 generic.go:334] "Generic (PLEG): container finished" podID="31e6220d-a6e0-4688-9827-7f889fd01fe8" containerID="18aef4d4eddeb58963317634637ac0624d83e58b8569eed0777394e52f173a60" exitCode=0 Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.648754 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" event={"ID":"31e6220d-a6e0-4688-9827-7f889fd01fe8","Type":"ContainerDied","Data":"18aef4d4eddeb58963317634637ac0624d83e58b8569eed0777394e52f173a60"} Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.649224 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" event={"ID":"31e6220d-a6e0-4688-9827-7f889fd01fe8","Type":"ContainerStarted","Data":"c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec"} Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.702667 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:01 crc kubenswrapper[4872]: I1009 08:45:01.751269 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:45:02 crc kubenswrapper[4872]: I1009 08:45:02.493060 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3231028e-ac1b-4f07-9bc6-3efb4160cfb0" path="/var/lib/kubelet/pods/3231028e-ac1b-4f07-9bc6-3efb4160cfb0/volumes" Oct 09 08:45:02 crc kubenswrapper[4872]: I1009 08:45:02.494250 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7407ecf1-4c1b-47cd-9c8c-099630211671" path="/var/lib/kubelet/pods/7407ecf1-4c1b-47cd-9c8c-099630211671/volumes" Oct 09 08:45:02 crc kubenswrapper[4872]: I1009 08:45:02.900730 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:02 crc kubenswrapper[4872]: I1009 08:45:02.900798 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:02 crc kubenswrapper[4872]: I1009 08:45:02.969452 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.149180 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume\") pod \"31e6220d-a6e0-4688-9827-7f889fd01fe8\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.149533 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rwpdq\" (UniqueName: \"kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq\") pod \"31e6220d-a6e0-4688-9827-7f889fd01fe8\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.149923 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume\") pod \"31e6220d-a6e0-4688-9827-7f889fd01fe8\" (UID: \"31e6220d-a6e0-4688-9827-7f889fd01fe8\") " Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.150953 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume" (OuterVolumeSpecName: "config-volume") pod "31e6220d-a6e0-4688-9827-7f889fd01fe8" (UID: "31e6220d-a6e0-4688-9827-7f889fd01fe8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.159364 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq" (OuterVolumeSpecName: "kube-api-access-rwpdq") pod "31e6220d-a6e0-4688-9827-7f889fd01fe8" (UID: "31e6220d-a6e0-4688-9827-7f889fd01fe8"). InnerVolumeSpecName "kube-api-access-rwpdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.161825 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "31e6220d-a6e0-4688-9827-7f889fd01fe8" (UID: "31e6220d-a6e0-4688-9827-7f889fd01fe8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.252181 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/31e6220d-a6e0-4688-9827-7f889fd01fe8-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.252220 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rwpdq\" (UniqueName: \"kubernetes.io/projected/31e6220d-a6e0-4688-9827-7f889fd01fe8-kube-api-access-rwpdq\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.252230 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/31e6220d-a6e0-4688-9827-7f889fd01fe8-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.672323 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ckl88" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="registry-server" containerID="cri-o://28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730" gracePeriod=2 Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.672998 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.673029 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333325-9zvjg" event={"ID":"31e6220d-a6e0-4688-9827-7f889fd01fe8","Type":"ContainerDied","Data":"c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec"} Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.673531 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f2850501b6125abcfa969679556ae0f7a65a3459429f33546bb3229fe9c4ec" Oct 09 08:45:03 crc kubenswrapper[4872]: I1009 08:45:03.953171 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k2jcv" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="registry-server" probeResult="failure" output=< Oct 09 08:45:03 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Oct 09 08:45:03 crc kubenswrapper[4872]: > Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.117950 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.270170 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities\") pod \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.270333 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content\") pod \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.270448 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m59qr\" (UniqueName: \"kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr\") pod \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\" (UID: \"97983f9f-be05-483b-b2f6-d532b1dfe8ea\") " Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.271017 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities" (OuterVolumeSpecName: "utilities") pod "97983f9f-be05-483b-b2f6-d532b1dfe8ea" (UID: "97983f9f-be05-483b-b2f6-d532b1dfe8ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.280009 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr" (OuterVolumeSpecName: "kube-api-access-m59qr") pod "97983f9f-be05-483b-b2f6-d532b1dfe8ea" (UID: "97983f9f-be05-483b-b2f6-d532b1dfe8ea"). InnerVolumeSpecName "kube-api-access-m59qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.282994 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97983f9f-be05-483b-b2f6-d532b1dfe8ea" (UID: "97983f9f-be05-483b-b2f6-d532b1dfe8ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.372311 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.372347 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97983f9f-be05-483b-b2f6-d532b1dfe8ea-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.372360 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m59qr\" (UniqueName: \"kubernetes.io/projected/97983f9f-be05-483b-b2f6-d532b1dfe8ea-kube-api-access-m59qr\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.686575 4872 generic.go:334] "Generic (PLEG): container finished" podID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerID="28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730" exitCode=0 Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.686622 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerDied","Data":"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730"} Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.686686 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ckl88" event={"ID":"97983f9f-be05-483b-b2f6-d532b1dfe8ea","Type":"ContainerDied","Data":"26946e53de4b2f00dd9bbbeb78ce449efe4044996a065a5f1150611be36801b1"} Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.686708 4872 scope.go:117] "RemoveContainer" containerID="28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.686751 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ckl88" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.715426 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.718996 4872 scope.go:117] "RemoveContainer" containerID="db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.725897 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ckl88"] Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.754013 4872 scope.go:117] "RemoveContainer" containerID="61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.792854 4872 scope.go:117] "RemoveContainer" containerID="28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730" Oct 09 08:45:04 crc kubenswrapper[4872]: E1009 08:45:04.793435 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730\": container with ID starting with 28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730 not found: ID does not exist" containerID="28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.793496 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730"} err="failed to get container status \"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730\": rpc error: code = NotFound desc = could not find container \"28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730\": container with ID starting with 28c4df856a8538d5d9feccef53c355c79dabafca28cde655b21db260aa060730 not found: ID does not exist" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.793531 4872 scope.go:117] "RemoveContainer" containerID="db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f" Oct 09 08:45:04 crc kubenswrapper[4872]: E1009 08:45:04.794003 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f\": container with ID starting with db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f not found: ID does not exist" containerID="db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.794046 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f"} err="failed to get container status \"db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f\": rpc error: code = NotFound desc = could not find container \"db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f\": container with ID starting with db5702a4a2693dd77ee9208aa8922367f64fb932fc3978042e9346908e3fd68f not found: ID does not exist" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.794075 4872 scope.go:117] "RemoveContainer" containerID="61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6" Oct 09 08:45:04 crc kubenswrapper[4872]: E1009 08:45:04.794316 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6\": container with ID starting with 61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6 not found: ID does not exist" containerID="61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6" Oct 09 08:45:04 crc kubenswrapper[4872]: I1009 08:45:04.794345 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6"} err="failed to get container status \"61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6\": rpc error: code = NotFound desc = could not find container \"61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6\": container with ID starting with 61b1862e7f57244da6c710806535a48f5626bd6646ee9e1d6217a6ce293044d6 not found: ID does not exist" Oct 09 08:45:05 crc kubenswrapper[4872]: I1009 08:45:05.462667 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:45:05 crc kubenswrapper[4872]: E1009 08:45:05.463249 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:45:06 crc kubenswrapper[4872]: I1009 08:45:06.489586 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" path="/var/lib/kubelet/pods/97983f9f-be05-483b-b2f6-d532b1dfe8ea/volumes" Oct 09 08:45:07 crc kubenswrapper[4872]: I1009 08:45:07.714997 4872 generic.go:334] "Generic (PLEG): container finished" podID="9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" containerID="e2f615a34f8c17349a4ebb82a435e035c2db8d4c55f22d126ec053b510c6d230" exitCode=0 Oct 09 08:45:07 crc kubenswrapper[4872]: I1009 08:45:07.715117 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" event={"ID":"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08","Type":"ContainerDied","Data":"e2f615a34f8c17349a4ebb82a435e035c2db8d4c55f22d126ec053b510c6d230"} Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.145089 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.265255 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key\") pod \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.265367 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s2jd2\" (UniqueName: \"kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2\") pod \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.265468 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory\") pod \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\" (UID: \"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08\") " Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.270866 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2" (OuterVolumeSpecName: "kube-api-access-s2jd2") pod "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" (UID: "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08"). InnerVolumeSpecName "kube-api-access-s2jd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.297335 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" (UID: "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.297361 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory" (OuterVolumeSpecName: "inventory") pod "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" (UID: "9b5aa1fc-253f-4c3b-b3f4-1364130a3a08"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.367865 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.367902 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s2jd2\" (UniqueName: \"kubernetes.io/projected/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-kube-api-access-s2jd2\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.367912 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9b5aa1fc-253f-4c3b-b3f4-1364130a3a08-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.737592 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" event={"ID":"9b5aa1fc-253f-4c3b-b3f4-1364130a3a08","Type":"ContainerDied","Data":"b3221a32e8de4f69b7d1554f6b78ec26e263bdb15b9265cad41478b9f71ad4d7"} Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.738261 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b3221a32e8de4f69b7d1554f6b78ec26e263bdb15b9265cad41478b9f71ad4d7" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.737663 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.814422 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg"] Oct 09 08:45:09 crc kubenswrapper[4872]: E1009 08:45:09.814957 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="extract-utilities" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.814980 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="extract-utilities" Oct 09 08:45:09 crc kubenswrapper[4872]: E1009 08:45:09.815007 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="extract-content" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815016 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="extract-content" Oct 09 08:45:09 crc kubenswrapper[4872]: E1009 08:45:09.815031 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815041 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 09 08:45:09 crc kubenswrapper[4872]: E1009 08:45:09.815061 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="registry-server" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815068 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="registry-server" Oct 09 08:45:09 crc kubenswrapper[4872]: E1009 08:45:09.815092 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31e6220d-a6e0-4688-9827-7f889fd01fe8" containerName="collect-profiles" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815101 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="31e6220d-a6e0-4688-9827-7f889fd01fe8" containerName="collect-profiles" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815323 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="97983f9f-be05-483b-b2f6-d532b1dfe8ea" containerName="registry-server" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815356 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b5aa1fc-253f-4c3b-b3f4-1364130a3a08" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.815383 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="31e6220d-a6e0-4688-9827-7f889fd01fe8" containerName="collect-profiles" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.816128 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.824803 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.824819 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.824865 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.825075 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.833581 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg"] Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.877361 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.877521 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4qjt\" (UniqueName: \"kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.877579 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.978434 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.978554 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.978627 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4qjt\" (UniqueName: \"kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.984477 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:09 crc kubenswrapper[4872]: I1009 08:45:09.984995 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:10 crc kubenswrapper[4872]: I1009 08:45:09.999944 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4qjt\" (UniqueName: \"kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:10 crc kubenswrapper[4872]: I1009 08:45:10.189201 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:45:10 crc kubenswrapper[4872]: I1009 08:45:10.721195 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg"] Oct 09 08:45:10 crc kubenswrapper[4872]: I1009 08:45:10.730011 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:45:10 crc kubenswrapper[4872]: I1009 08:45:10.747012 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" event={"ID":"57a6c64a-020d-4b90-b350-84f9441d08b3","Type":"ContainerStarted","Data":"d6358f05f59b114f8b5ddd7c723691926a57e4151fc9d6575dfbeddeb67a165f"} Oct 09 08:45:11 crc kubenswrapper[4872]: I1009 08:45:11.769271 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" event={"ID":"57a6c64a-020d-4b90-b350-84f9441d08b3","Type":"ContainerStarted","Data":"9735fe880763c78edb2d29567ee578e43c0751408ae17d20b2c8f84ef12f96c7"} Oct 09 08:45:11 crc kubenswrapper[4872]: I1009 08:45:11.792928 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" podStartSLOduration=2.219599556 podStartE2EDuration="2.79290601s" podCreationTimestamp="2025-10-09 08:45:09 +0000 UTC" firstStartedPulling="2025-10-09 08:45:10.729808784 +0000 UTC m=+1668.920337410" lastFinishedPulling="2025-10-09 08:45:11.303115238 +0000 UTC m=+1669.493643864" observedRunningTime="2025-10-09 08:45:11.785068252 +0000 UTC m=+1669.975596888" watchObservedRunningTime="2025-10-09 08:45:11.79290601 +0000 UTC m=+1669.983434646" Oct 09 08:45:12 crc kubenswrapper[4872]: I1009 08:45:12.950118 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:12 crc kubenswrapper[4872]: I1009 08:45:12.996248 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.050074 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-8678-account-create-xs9w6"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.064654 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-ce3d-account-create-65snb"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.067553 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-cef4-account-create-nr247"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.076680 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-8678-account-create-xs9w6"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.084745 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-ce3d-account-create-65snb"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.092452 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-cef4-account-create-nr247"] Oct 09 08:45:13 crc kubenswrapper[4872]: I1009 08:45:13.186846 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:45:14 crc kubenswrapper[4872]: I1009 08:45:14.473469 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c611709-a037-41b7-a8f6-1dc9fdbb8d4d" path="/var/lib/kubelet/pods/1c611709-a037-41b7-a8f6-1dc9fdbb8d4d/volumes" Oct 09 08:45:14 crc kubenswrapper[4872]: I1009 08:45:14.474711 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3032b0c-ddae-450b-be63-1cd4b9014afb" path="/var/lib/kubelet/pods/b3032b0c-ddae-450b-be63-1cd4b9014afb/volumes" Oct 09 08:45:14 crc kubenswrapper[4872]: I1009 08:45:14.475807 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6154eba-fb06-4583-8937-9c1183083fec" path="/var/lib/kubelet/pods/b6154eba-fb06-4583-8937-9c1183083fec/volumes" Oct 09 08:45:14 crc kubenswrapper[4872]: I1009 08:45:14.798486 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k2jcv" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="registry-server" containerID="cri-o://cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458" gracePeriod=2 Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.035495 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-wwgjr"] Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.045326 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-wwgjr"] Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.264268 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.317762 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptzcq\" (UniqueName: \"kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq\") pod \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.318000 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content\") pod \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.318177 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities\") pod \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\" (UID: \"2d3ecaf9-e22b-4385-be29-a1340d30cb54\") " Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.318925 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities" (OuterVolumeSpecName: "utilities") pod "2d3ecaf9-e22b-4385-be29-a1340d30cb54" (UID: "2d3ecaf9-e22b-4385-be29-a1340d30cb54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.324480 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq" (OuterVolumeSpecName: "kube-api-access-ptzcq") pod "2d3ecaf9-e22b-4385-be29-a1340d30cb54" (UID: "2d3ecaf9-e22b-4385-be29-a1340d30cb54"). InnerVolumeSpecName "kube-api-access-ptzcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.390164 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d3ecaf9-e22b-4385-be29-a1340d30cb54" (UID: "2d3ecaf9-e22b-4385-be29-a1340d30cb54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.420115 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.420157 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptzcq\" (UniqueName: \"kubernetes.io/projected/2d3ecaf9-e22b-4385-be29-a1340d30cb54-kube-api-access-ptzcq\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.420169 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d3ecaf9-e22b-4385-be29-a1340d30cb54-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.813284 4872 generic.go:334] "Generic (PLEG): container finished" podID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerID="cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458" exitCode=0 Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.813325 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerDied","Data":"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458"} Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.813352 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k2jcv" event={"ID":"2d3ecaf9-e22b-4385-be29-a1340d30cb54","Type":"ContainerDied","Data":"ff10600965f4d52420b7475a8a1ac528c9984a55a8829c99df9095daa48d3a6e"} Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.813370 4872 scope.go:117] "RemoveContainer" containerID="cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.813383 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k2jcv" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.846235 4872 scope.go:117] "RemoveContainer" containerID="51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.858437 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.865599 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k2jcv"] Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.888281 4872 scope.go:117] "RemoveContainer" containerID="5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.925429 4872 scope.go:117] "RemoveContainer" containerID="cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458" Oct 09 08:45:15 crc kubenswrapper[4872]: E1009 08:45:15.925918 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458\": container with ID starting with cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458 not found: ID does not exist" containerID="cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.925959 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458"} err="failed to get container status \"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458\": rpc error: code = NotFound desc = could not find container \"cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458\": container with ID starting with cf1ad70342ecb95610349b275c6af893d3eaa0d273a49d1ba2ed948916f2a458 not found: ID does not exist" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.925986 4872 scope.go:117] "RemoveContainer" containerID="51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88" Oct 09 08:45:15 crc kubenswrapper[4872]: E1009 08:45:15.926316 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88\": container with ID starting with 51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88 not found: ID does not exist" containerID="51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.926347 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88"} err="failed to get container status \"51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88\": rpc error: code = NotFound desc = could not find container \"51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88\": container with ID starting with 51c763b06df5d226e82d24a20ac6ac0631c1e9407d8cde15a93d8c24e8122b88 not found: ID does not exist" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.926364 4872 scope.go:117] "RemoveContainer" containerID="5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a" Oct 09 08:45:15 crc kubenswrapper[4872]: E1009 08:45:15.926800 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a\": container with ID starting with 5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a not found: ID does not exist" containerID="5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a" Oct 09 08:45:15 crc kubenswrapper[4872]: I1009 08:45:15.926828 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a"} err="failed to get container status \"5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a\": rpc error: code = NotFound desc = could not find container \"5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a\": container with ID starting with 5b238f19c49e8250019929cf50ae1ef04e702a392fe625481c3993861c03999a not found: ID does not exist" Oct 09 08:45:16 crc kubenswrapper[4872]: I1009 08:45:16.461509 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:45:16 crc kubenswrapper[4872]: E1009 08:45:16.461818 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:45:16 crc kubenswrapper[4872]: I1009 08:45:16.477581 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" path="/var/lib/kubelet/pods/2d3ecaf9-e22b-4385-be29-a1340d30cb54/volumes" Oct 09 08:45:16 crc kubenswrapper[4872]: I1009 08:45:16.478507 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5134c98-fe5f-4fba-ac37-e3c32b0ba26c" path="/var/lib/kubelet/pods/f5134c98-fe5f-4fba-ac37-e3c32b0ba26c/volumes" Oct 09 08:45:30 crc kubenswrapper[4872]: I1009 08:45:30.462151 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:45:30 crc kubenswrapper[4872]: E1009 08:45:30.462957 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:45:34 crc kubenswrapper[4872]: I1009 08:45:34.028928 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-jvtzz"] Oct 09 08:45:34 crc kubenswrapper[4872]: I1009 08:45:34.036882 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-jvtzz"] Oct 09 08:45:34 crc kubenswrapper[4872]: I1009 08:45:34.486818 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="256fd449-62bd-44fe-945e-92663ab62698" path="/var/lib/kubelet/pods/256fd449-62bd-44fe-945e-92663ab62698/volumes" Oct 09 08:45:45 crc kubenswrapper[4872]: I1009 08:45:45.461365 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:45:45 crc kubenswrapper[4872]: E1009 08:45:45.462155 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.217092 4872 scope.go:117] "RemoveContainer" containerID="681e731370014a39b6a07c191b12504f3a40e0d0ff39f36f69cd2a0c1fecfc7c" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.246485 4872 scope.go:117] "RemoveContainer" containerID="ce0c4c5c5b53ea1f12864087e6d9a473e854d32244f134fe3f0db4ca61e58a02" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.271544 4872 scope.go:117] "RemoveContainer" containerID="3a64f434640ff8cb81b37264daa7527ee8f026aff3b786d914700dc7fbfe50bd" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.327282 4872 scope.go:117] "RemoveContainer" containerID="fc5356d5eb0b02f29e7d0e03b74c7bf865150c941377f1311308f3be7e10c087" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.409812 4872 scope.go:117] "RemoveContainer" containerID="8a0c592c04250cdba12ebb992b64f712f9b5e1aefe15e21c76ea554cac59e51e" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.433034 4872 scope.go:117] "RemoveContainer" containerID="12437e4f66176fed9155f6ea28e39b7f9b4efd6613206c6839dfa04e52d476b3" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.480675 4872 scope.go:117] "RemoveContainer" containerID="84e6959f826be17cb56ea1b113ca20739fa8c18b27e7ee8b3e9e2559c7d093b8" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.534812 4872 scope.go:117] "RemoveContainer" containerID="5085d9b21dfba4b5228d6fa27fd7273de62f29732f8645437c5109b0d928d618" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.554065 4872 scope.go:117] "RemoveContainer" containerID="045016146c016ce9c181dfc7b589775fb105549bf098f7943aeda130b5eaf38f" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.582088 4872 scope.go:117] "RemoveContainer" containerID="f4b32ad768096201576d9394662a22fb88db28900c4212fa60dd91a498d8bc6b" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.599108 4872 scope.go:117] "RemoveContainer" containerID="c68ade29cc29c658f9e26ea63dce63dea34e96435930126fdaf774d09d87827b" Oct 09 08:45:46 crc kubenswrapper[4872]: I1009 08:45:46.619420 4872 scope.go:117] "RemoveContainer" containerID="c7110c0588155b4a859531ff90cdb9135da5fb43cbd6f96f5902fee1c34f1fbe" Oct 09 08:45:51 crc kubenswrapper[4872]: I1009 08:45:51.039352 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-gbw26"] Oct 09 08:45:51 crc kubenswrapper[4872]: I1009 08:45:51.050461 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-b9jgn"] Oct 09 08:45:51 crc kubenswrapper[4872]: I1009 08:45:51.059682 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-gbw26"] Oct 09 08:45:51 crc kubenswrapper[4872]: I1009 08:45:51.069184 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-b9jgn"] Oct 09 08:45:52 crc kubenswrapper[4872]: I1009 08:45:52.475241 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b8d1d3f-8d6e-458f-8ebe-da41817d436c" path="/var/lib/kubelet/pods/3b8d1d3f-8d6e-458f-8ebe-da41817d436c/volumes" Oct 09 08:45:52 crc kubenswrapper[4872]: I1009 08:45:52.477161 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6788083-19e2-4636-9c90-468ddecb1115" path="/var/lib/kubelet/pods/c6788083-19e2-4636-9c90-468ddecb1115/volumes" Oct 09 08:45:57 crc kubenswrapper[4872]: I1009 08:45:57.481105 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:45:57 crc kubenswrapper[4872]: E1009 08:45:57.484550 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:46:00 crc kubenswrapper[4872]: I1009 08:46:00.036830 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zprz2"] Oct 09 08:46:00 crc kubenswrapper[4872]: I1009 08:46:00.045132 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zprz2"] Oct 09 08:46:00 crc kubenswrapper[4872]: I1009 08:46:00.483596 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1376b2ed-ee38-4201-bfd3-f35ef4093fa0" path="/var/lib/kubelet/pods/1376b2ed-ee38-4201-bfd3-f35ef4093fa0/volumes" Oct 09 08:46:11 crc kubenswrapper[4872]: I1009 08:46:11.045072 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-j52s6"] Oct 09 08:46:11 crc kubenswrapper[4872]: I1009 08:46:11.055426 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-j52s6"] Oct 09 08:46:12 crc kubenswrapper[4872]: I1009 08:46:12.468921 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:46:12 crc kubenswrapper[4872]: E1009 08:46:12.469360 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:46:12 crc kubenswrapper[4872]: I1009 08:46:12.472297 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658ddbb1-1aaf-4e34-9a59-c6d01a87c505" path="/var/lib/kubelet/pods/658ddbb1-1aaf-4e34-9a59-c6d01a87c505/volumes" Oct 09 08:46:13 crc kubenswrapper[4872]: I1009 08:46:13.026252 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-8jxqn"] Oct 09 08:46:13 crc kubenswrapper[4872]: I1009 08:46:13.036512 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-8jxqn"] Oct 09 08:46:14 crc kubenswrapper[4872]: I1009 08:46:14.472152 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88cf1623-86ec-41eb-802f-c0996b7442be" path="/var/lib/kubelet/pods/88cf1623-86ec-41eb-802f-c0996b7442be/volumes" Oct 09 08:46:23 crc kubenswrapper[4872]: I1009 08:46:23.473620 4872 generic.go:334] "Generic (PLEG): container finished" podID="57a6c64a-020d-4b90-b350-84f9441d08b3" containerID="9735fe880763c78edb2d29567ee578e43c0751408ae17d20b2c8f84ef12f96c7" exitCode=0 Oct 09 08:46:23 crc kubenswrapper[4872]: I1009 08:46:23.473695 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" event={"ID":"57a6c64a-020d-4b90-b350-84f9441d08b3","Type":"ContainerDied","Data":"9735fe880763c78edb2d29567ee578e43c0751408ae17d20b2c8f84ef12f96c7"} Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.461356 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:46:24 crc kubenswrapper[4872]: E1009 08:46:24.461867 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.867370 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.943998 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory\") pod \"57a6c64a-020d-4b90-b350-84f9441d08b3\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.944099 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key\") pod \"57a6c64a-020d-4b90-b350-84f9441d08b3\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.944324 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4qjt\" (UniqueName: \"kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt\") pod \"57a6c64a-020d-4b90-b350-84f9441d08b3\" (UID: \"57a6c64a-020d-4b90-b350-84f9441d08b3\") " Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.950882 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt" (OuterVolumeSpecName: "kube-api-access-l4qjt") pod "57a6c64a-020d-4b90-b350-84f9441d08b3" (UID: "57a6c64a-020d-4b90-b350-84f9441d08b3"). InnerVolumeSpecName "kube-api-access-l4qjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.975099 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory" (OuterVolumeSpecName: "inventory") pod "57a6c64a-020d-4b90-b350-84f9441d08b3" (UID: "57a6c64a-020d-4b90-b350-84f9441d08b3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:46:24 crc kubenswrapper[4872]: I1009 08:46:24.977041 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "57a6c64a-020d-4b90-b350-84f9441d08b3" (UID: "57a6c64a-020d-4b90-b350-84f9441d08b3"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.046131 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4qjt\" (UniqueName: \"kubernetes.io/projected/57a6c64a-020d-4b90-b350-84f9441d08b3-kube-api-access-l4qjt\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.046170 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.046183 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/57a6c64a-020d-4b90-b350-84f9441d08b3-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.493363 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" event={"ID":"57a6c64a-020d-4b90-b350-84f9441d08b3","Type":"ContainerDied","Data":"d6358f05f59b114f8b5ddd7c723691926a57e4151fc9d6575dfbeddeb67a165f"} Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.493416 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6358f05f59b114f8b5ddd7c723691926a57e4151fc9d6575dfbeddeb67a165f" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.493444 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.574820 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc"] Oct 09 08:46:25 crc kubenswrapper[4872]: E1009 08:46:25.575643 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57a6c64a-020d-4b90-b350-84f9441d08b3" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575684 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="57a6c64a-020d-4b90-b350-84f9441d08b3" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:25 crc kubenswrapper[4872]: E1009 08:46:25.575704 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="extract-content" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575711 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="extract-content" Oct 09 08:46:25 crc kubenswrapper[4872]: E1009 08:46:25.575722 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="registry-server" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575727 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="registry-server" Oct 09 08:46:25 crc kubenswrapper[4872]: E1009 08:46:25.575737 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="extract-utilities" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575743 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="extract-utilities" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575915 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="57a6c64a-020d-4b90-b350-84f9441d08b3" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.575933 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d3ecaf9-e22b-4385-be29-a1340d30cb54" containerName="registry-server" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.577768 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.580199 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.581553 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.581789 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.583035 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.599089 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc"] Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.656688 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.656817 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.656947 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5qrg\" (UniqueName: \"kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.759475 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.759577 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.759681 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5qrg\" (UniqueName: \"kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.764400 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.764576 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.777175 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5qrg\" (UniqueName: \"kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-msgrc\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:25 crc kubenswrapper[4872]: I1009 08:46:25.898058 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:26 crc kubenswrapper[4872]: I1009 08:46:26.408024 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc"] Oct 09 08:46:26 crc kubenswrapper[4872]: W1009 08:46:26.414868 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3586ab35_c3cc_4f05_8319_b0b568c2c029.slice/crio-60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62 WatchSource:0}: Error finding container 60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62: Status 404 returned error can't find the container with id 60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62 Oct 09 08:46:26 crc kubenswrapper[4872]: I1009 08:46:26.503942 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" event={"ID":"3586ab35-c3cc-4f05-8319-b0b568c2c029","Type":"ContainerStarted","Data":"60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62"} Oct 09 08:46:27 crc kubenswrapper[4872]: I1009 08:46:27.517933 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" event={"ID":"3586ab35-c3cc-4f05-8319-b0b568c2c029","Type":"ContainerStarted","Data":"1155acf1ec58be1bdffa8cf5399710d8c29f32a6942f51837b8625da741dde1a"} Oct 09 08:46:27 crc kubenswrapper[4872]: I1009 08:46:27.535552 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" podStartSLOduration=2.040017277 podStartE2EDuration="2.535532635s" podCreationTimestamp="2025-10-09 08:46:25 +0000 UTC" firstStartedPulling="2025-10-09 08:46:26.417896024 +0000 UTC m=+1744.608424650" lastFinishedPulling="2025-10-09 08:46:26.913411382 +0000 UTC m=+1745.103940008" observedRunningTime="2025-10-09 08:46:27.533487666 +0000 UTC m=+1745.724016312" watchObservedRunningTime="2025-10-09 08:46:27.535532635 +0000 UTC m=+1745.726061261" Oct 09 08:46:31 crc kubenswrapper[4872]: I1009 08:46:31.567447 4872 generic.go:334] "Generic (PLEG): container finished" podID="3586ab35-c3cc-4f05-8319-b0b568c2c029" containerID="1155acf1ec58be1bdffa8cf5399710d8c29f32a6942f51837b8625da741dde1a" exitCode=0 Oct 09 08:46:31 crc kubenswrapper[4872]: I1009 08:46:31.567562 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" event={"ID":"3586ab35-c3cc-4f05-8319-b0b568c2c029","Type":"ContainerDied","Data":"1155acf1ec58be1bdffa8cf5399710d8c29f32a6942f51837b8625da741dde1a"} Oct 09 08:46:32 crc kubenswrapper[4872]: I1009 08:46:32.974584 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.012204 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory\") pod \"3586ab35-c3cc-4f05-8319-b0b568c2c029\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.012261 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5qrg\" (UniqueName: \"kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg\") pod \"3586ab35-c3cc-4f05-8319-b0b568c2c029\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.012291 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key\") pod \"3586ab35-c3cc-4f05-8319-b0b568c2c029\" (UID: \"3586ab35-c3cc-4f05-8319-b0b568c2c029\") " Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.018942 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg" (OuterVolumeSpecName: "kube-api-access-f5qrg") pod "3586ab35-c3cc-4f05-8319-b0b568c2c029" (UID: "3586ab35-c3cc-4f05-8319-b0b568c2c029"). InnerVolumeSpecName "kube-api-access-f5qrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.040337 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory" (OuterVolumeSpecName: "inventory") pod "3586ab35-c3cc-4f05-8319-b0b568c2c029" (UID: "3586ab35-c3cc-4f05-8319-b0b568c2c029"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.042625 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3586ab35-c3cc-4f05-8319-b0b568c2c029" (UID: "3586ab35-c3cc-4f05-8319-b0b568c2c029"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.115804 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.115850 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3586ab35-c3cc-4f05-8319-b0b568c2c029-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.115864 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5qrg\" (UniqueName: \"kubernetes.io/projected/3586ab35-c3cc-4f05-8319-b0b568c2c029-kube-api-access-f5qrg\") on node \"crc\" DevicePath \"\"" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.587455 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" event={"ID":"3586ab35-c3cc-4f05-8319-b0b568c2c029","Type":"ContainerDied","Data":"60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62"} Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.587821 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60085f70877e47eb2dd1c5abff66393f908196e4de68c2fa2280cc9db7c9ec62" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.587520 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-msgrc" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.708785 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4"] Oct 09 08:46:33 crc kubenswrapper[4872]: E1009 08:46:33.709251 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3586ab35-c3cc-4f05-8319-b0b568c2c029" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.709274 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3586ab35-c3cc-4f05-8319-b0b568c2c029" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.709509 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3586ab35-c3cc-4f05-8319-b0b568c2c029" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.710345 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.715720 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.715833 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.718331 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.718736 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.736121 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4"] Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.746411 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.746519 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ttcq\" (UniqueName: \"kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.746724 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.847763 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.847853 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ttcq\" (UniqueName: \"kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.847932 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.853249 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.869550 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ttcq\" (UniqueName: \"kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:33 crc kubenswrapper[4872]: I1009 08:46:33.870225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qdjb4\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:34 crc kubenswrapper[4872]: I1009 08:46:34.053430 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:46:34 crc kubenswrapper[4872]: I1009 08:46:34.569678 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4"] Oct 09 08:46:34 crc kubenswrapper[4872]: I1009 08:46:34.599018 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" event={"ID":"eb153425-cc47-4785-be11-49785db1cca7","Type":"ContainerStarted","Data":"88b4b4468705e0be0b98f9f61aa13bec4a74b700b6f56f646b7bfb83bef2ae15"} Oct 09 08:46:36 crc kubenswrapper[4872]: I1009 08:46:36.623423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" event={"ID":"eb153425-cc47-4785-be11-49785db1cca7","Type":"ContainerStarted","Data":"67a77c2869bc6849aee40f48496bbc590d946752c590d14afa5d7c64301bb704"} Oct 09 08:46:36 crc kubenswrapper[4872]: I1009 08:46:36.654264 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" podStartSLOduration=2.783917051 podStartE2EDuration="3.654244551s" podCreationTimestamp="2025-10-09 08:46:33 +0000 UTC" firstStartedPulling="2025-10-09 08:46:34.573301358 +0000 UTC m=+1752.763829984" lastFinishedPulling="2025-10-09 08:46:35.443628818 +0000 UTC m=+1753.634157484" observedRunningTime="2025-10-09 08:46:36.641081318 +0000 UTC m=+1754.831609964" watchObservedRunningTime="2025-10-09 08:46:36.654244551 +0000 UTC m=+1754.844773177" Oct 09 08:46:39 crc kubenswrapper[4872]: I1009 08:46:39.462157 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:46:39 crc kubenswrapper[4872]: E1009 08:46:39.462815 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:46:46 crc kubenswrapper[4872]: I1009 08:46:46.842421 4872 scope.go:117] "RemoveContainer" containerID="b5298ecc9b22f9956f047d1fc258e0013e0ade9f95007721882a668864ed57f7" Oct 09 08:46:46 crc kubenswrapper[4872]: I1009 08:46:46.884390 4872 scope.go:117] "RemoveContainer" containerID="0b4f7e667c969fd1c95e3e6f9770ccc34baf56a429957df46173bbf0b399d703" Oct 09 08:46:46 crc kubenswrapper[4872]: I1009 08:46:46.938854 4872 scope.go:117] "RemoveContainer" containerID="f8f4f178e43c02d36259b6374f52177a6612221341d63c993728d95100aa152d" Oct 09 08:46:46 crc kubenswrapper[4872]: I1009 08:46:46.979281 4872 scope.go:117] "RemoveContainer" containerID="9f6d857c409a7c3371ea3f396ae5b825dab62973f557011b10fb55d0bf935d99" Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.015183 4872 scope.go:117] "RemoveContainer" containerID="b2f2219e4942c73b94cb762a64e5ee5f127a4f829ea5c155feeb42e570e44b3e" Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.044661 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-kl2lk"] Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.066273 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-wcspf"] Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.077350 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-dbs6x"] Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.085972 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-kl2lk"] Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.093619 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-dbs6x"] Oct 09 08:46:47 crc kubenswrapper[4872]: I1009 08:46:47.100331 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-wcspf"] Oct 09 08:46:48 crc kubenswrapper[4872]: I1009 08:46:48.474784 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="896adf7e-0f68-4f60-bd0c-6e507673c221" path="/var/lib/kubelet/pods/896adf7e-0f68-4f60-bd0c-6e507673c221/volumes" Oct 09 08:46:48 crc kubenswrapper[4872]: I1009 08:46:48.476015 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0703df6-2aa8-4de5-b54e-850b63effded" path="/var/lib/kubelet/pods/a0703df6-2aa8-4de5-b54e-850b63effded/volumes" Oct 09 08:46:48 crc kubenswrapper[4872]: I1009 08:46:48.477015 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aed630ee-8d81-4c05-8cd0-85245b99fa00" path="/var/lib/kubelet/pods/aed630ee-8d81-4c05-8cd0-85245b99fa00/volumes" Oct 09 08:46:50 crc kubenswrapper[4872]: I1009 08:46:50.461802 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:46:50 crc kubenswrapper[4872]: E1009 08:46:50.462085 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.025345 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-1b9f-account-create-k5422"] Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.033467 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9786-account-create-dz7zk"] Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.042370 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-1b9f-account-create-k5422"] Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.050307 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9786-account-create-dz7zk"] Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.472497 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0531d081-32b1-4166-9134-cfc92932b17f" path="/var/lib/kubelet/pods/0531d081-32b1-4166-9134-cfc92932b17f/volumes" Oct 09 08:47:02 crc kubenswrapper[4872]: I1009 08:47:02.473092 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a78072f1-019e-478e-8736-c267e5604d4d" path="/var/lib/kubelet/pods/a78072f1-019e-478e-8736-c267e5604d4d/volumes" Oct 09 08:47:03 crc kubenswrapper[4872]: I1009 08:47:03.033539 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-051f-account-create-4mkf8"] Oct 09 08:47:03 crc kubenswrapper[4872]: I1009 08:47:03.042289 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-051f-account-create-4mkf8"] Oct 09 08:47:03 crc kubenswrapper[4872]: I1009 08:47:03.461681 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:47:03 crc kubenswrapper[4872]: E1009 08:47:03.462071 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:04 crc kubenswrapper[4872]: I1009 08:47:04.473445 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e640feb9-e26f-45e3-9e6b-06be8678076e" path="/var/lib/kubelet/pods/e640feb9-e26f-45e3-9e6b-06be8678076e/volumes" Oct 09 08:47:10 crc kubenswrapper[4872]: I1009 08:47:10.934415 4872 generic.go:334] "Generic (PLEG): container finished" podID="eb153425-cc47-4785-be11-49785db1cca7" containerID="67a77c2869bc6849aee40f48496bbc590d946752c590d14afa5d7c64301bb704" exitCode=0 Oct 09 08:47:10 crc kubenswrapper[4872]: I1009 08:47:10.934527 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" event={"ID":"eb153425-cc47-4785-be11-49785db1cca7","Type":"ContainerDied","Data":"67a77c2869bc6849aee40f48496bbc590d946752c590d14afa5d7c64301bb704"} Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.351288 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.483754 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key\") pod \"eb153425-cc47-4785-be11-49785db1cca7\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.484097 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory\") pod \"eb153425-cc47-4785-be11-49785db1cca7\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.484221 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ttcq\" (UniqueName: \"kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq\") pod \"eb153425-cc47-4785-be11-49785db1cca7\" (UID: \"eb153425-cc47-4785-be11-49785db1cca7\") " Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.489226 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq" (OuterVolumeSpecName: "kube-api-access-5ttcq") pod "eb153425-cc47-4785-be11-49785db1cca7" (UID: "eb153425-cc47-4785-be11-49785db1cca7"). InnerVolumeSpecName "kube-api-access-5ttcq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.514397 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory" (OuterVolumeSpecName: "inventory") pod "eb153425-cc47-4785-be11-49785db1cca7" (UID: "eb153425-cc47-4785-be11-49785db1cca7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.518734 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "eb153425-cc47-4785-be11-49785db1cca7" (UID: "eb153425-cc47-4785-be11-49785db1cca7"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.586920 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ttcq\" (UniqueName: \"kubernetes.io/projected/eb153425-cc47-4785-be11-49785db1cca7-kube-api-access-5ttcq\") on node \"crc\" DevicePath \"\"" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.586958 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.586970 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/eb153425-cc47-4785-be11-49785db1cca7-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.952035 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" event={"ID":"eb153425-cc47-4785-be11-49785db1cca7","Type":"ContainerDied","Data":"88b4b4468705e0be0b98f9f61aa13bec4a74b700b6f56f646b7bfb83bef2ae15"} Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.952353 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b4b4468705e0be0b98f9f61aa13bec4a74b700b6f56f646b7bfb83bef2ae15" Oct 09 08:47:12 crc kubenswrapper[4872]: I1009 08:47:12.952088 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qdjb4" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.041842 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq"] Oct 09 08:47:13 crc kubenswrapper[4872]: E1009 08:47:13.042275 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb153425-cc47-4785-be11-49785db1cca7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.045174 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb153425-cc47-4785-be11-49785db1cca7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.046786 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb153425-cc47-4785-be11-49785db1cca7" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.047513 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.049816 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.050161 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.050230 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.050621 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.055646 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq"] Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.096435 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf6xl\" (UniqueName: \"kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.096503 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.096533 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.198692 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.198956 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf6xl\" (UniqueName: \"kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.199021 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.203225 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.204171 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.218313 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf6xl\" (UniqueName: \"kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-f24mq\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.370430 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.865730 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq"] Oct 09 08:47:13 crc kubenswrapper[4872]: I1009 08:47:13.961134 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" event={"ID":"f0febe64-b028-4782-a3e8-31b07031c230","Type":"ContainerStarted","Data":"2c89892086555bf8b0e90ce9438e21f2f4f1babbda7b78ce6dcaf349eda1278b"} Oct 09 08:47:14 crc kubenswrapper[4872]: I1009 08:47:14.462134 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:47:14 crc kubenswrapper[4872]: E1009 08:47:14.462787 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:14 crc kubenswrapper[4872]: I1009 08:47:14.983862 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" event={"ID":"f0febe64-b028-4782-a3e8-31b07031c230","Type":"ContainerStarted","Data":"b8181c5ff664bdb6f7fd7dd3c03b697d16a426cce2835142f3e1fa54041a8e3e"} Oct 09 08:47:15 crc kubenswrapper[4872]: I1009 08:47:15.003733 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" podStartSLOduration=1.597896507 podStartE2EDuration="2.0037148s" podCreationTimestamp="2025-10-09 08:47:13 +0000 UTC" firstStartedPulling="2025-10-09 08:47:13.870584359 +0000 UTC m=+1792.061112985" lastFinishedPulling="2025-10-09 08:47:14.276402652 +0000 UTC m=+1792.466931278" observedRunningTime="2025-10-09 08:47:14.998165399 +0000 UTC m=+1793.188694025" watchObservedRunningTime="2025-10-09 08:47:15.0037148 +0000 UTC m=+1793.194243426" Oct 09 08:47:27 crc kubenswrapper[4872]: I1009 08:47:27.462157 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:47:27 crc kubenswrapper[4872]: E1009 08:47:27.463720 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:28 crc kubenswrapper[4872]: I1009 08:47:28.055118 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b7nfs"] Oct 09 08:47:28 crc kubenswrapper[4872]: I1009 08:47:28.066180 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b7nfs"] Oct 09 08:47:28 crc kubenswrapper[4872]: I1009 08:47:28.475873 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="968a20fd-dd60-49f3-8e98-0d7b9e059328" path="/var/lib/kubelet/pods/968a20fd-dd60-49f3-8e98-0d7b9e059328/volumes" Oct 09 08:47:39 crc kubenswrapper[4872]: I1009 08:47:39.462324 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:47:39 crc kubenswrapper[4872]: E1009 08:47:39.464017 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.154876 4872 scope.go:117] "RemoveContainer" containerID="7a2c046245acf6e146feb206c3b9b6c773233f5af097f5f281683cc399a43f08" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.178398 4872 scope.go:117] "RemoveContainer" containerID="2ac02529dc578dc33ef21aca83ed8fbc82b9be1573c450ae5147e58eb5c37167" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.237538 4872 scope.go:117] "RemoveContainer" containerID="b1066cfed7505454036e5cc7ea38daf5d26a5c1bdb956384641b90fbd2390dea" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.289435 4872 scope.go:117] "RemoveContainer" containerID="6de5737ca4cd4443921fe5b3136243b0206ba4c3da09e6a607757b640233d327" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.338564 4872 scope.go:117] "RemoveContainer" containerID="823e451745a47d839ae0ddb052a09bbc3b7474a2b065da9f764e0bc14ae702da" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.378149 4872 scope.go:117] "RemoveContainer" containerID="e7997ef6320771fffc338d4411410a8304acf25cf8979aff8fd22d179dff1e5e" Oct 09 08:47:47 crc kubenswrapper[4872]: I1009 08:47:47.418323 4872 scope.go:117] "RemoveContainer" containerID="9b57439369b1d66ba5273fd53bf568d3cdc851b60932530e41a5606ac41aed04" Oct 09 08:47:50 crc kubenswrapper[4872]: I1009 08:47:50.041813 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-htgks"] Oct 09 08:47:50 crc kubenswrapper[4872]: I1009 08:47:50.051228 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-htgks"] Oct 09 08:47:50 crc kubenswrapper[4872]: I1009 08:47:50.476323 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96070559-6af7-4827-8dd7-38edfdd2ac47" path="/var/lib/kubelet/pods/96070559-6af7-4827-8dd7-38edfdd2ac47/volumes" Oct 09 08:47:51 crc kubenswrapper[4872]: I1009 08:47:51.461891 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:47:51 crc kubenswrapper[4872]: E1009 08:47:51.462212 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:47:52 crc kubenswrapper[4872]: I1009 08:47:52.031963 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hrvkd"] Oct 09 08:47:52 crc kubenswrapper[4872]: I1009 08:47:52.038903 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-hrvkd"] Oct 09 08:47:52 crc kubenswrapper[4872]: I1009 08:47:52.473226 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b76157d-7dc5-4661-972e-6673f3bd9990" path="/var/lib/kubelet/pods/3b76157d-7dc5-4661-972e-6673f3bd9990/volumes" Oct 09 08:48:06 crc kubenswrapper[4872]: I1009 08:48:06.462514 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:48:06 crc kubenswrapper[4872]: E1009 08:48:06.463398 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:48:07 crc kubenswrapper[4872]: I1009 08:48:07.453389 4872 generic.go:334] "Generic (PLEG): container finished" podID="f0febe64-b028-4782-a3e8-31b07031c230" containerID="b8181c5ff664bdb6f7fd7dd3c03b697d16a426cce2835142f3e1fa54041a8e3e" exitCode=2 Oct 09 08:48:07 crc kubenswrapper[4872]: I1009 08:48:07.453533 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" event={"ID":"f0febe64-b028-4782-a3e8-31b07031c230","Type":"ContainerDied","Data":"b8181c5ff664bdb6f7fd7dd3c03b697d16a426cce2835142f3e1fa54041a8e3e"} Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.869428 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.872495 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf6xl\" (UniqueName: \"kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl\") pod \"f0febe64-b028-4782-a3e8-31b07031c230\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.872702 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory\") pod \"f0febe64-b028-4782-a3e8-31b07031c230\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.872839 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key\") pod \"f0febe64-b028-4782-a3e8-31b07031c230\" (UID: \"f0febe64-b028-4782-a3e8-31b07031c230\") " Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.878859 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl" (OuterVolumeSpecName: "kube-api-access-mf6xl") pod "f0febe64-b028-4782-a3e8-31b07031c230" (UID: "f0febe64-b028-4782-a3e8-31b07031c230"). InnerVolumeSpecName "kube-api-access-mf6xl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.900916 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f0febe64-b028-4782-a3e8-31b07031c230" (UID: "f0febe64-b028-4782-a3e8-31b07031c230"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.919762 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory" (OuterVolumeSpecName: "inventory") pod "f0febe64-b028-4782-a3e8-31b07031c230" (UID: "f0febe64-b028-4782-a3e8-31b07031c230"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.975237 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.975290 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf6xl\" (UniqueName: \"kubernetes.io/projected/f0febe64-b028-4782-a3e8-31b07031c230-kube-api-access-mf6xl\") on node \"crc\" DevicePath \"\"" Oct 09 08:48:08 crc kubenswrapper[4872]: I1009 08:48:08.975308 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f0febe64-b028-4782-a3e8-31b07031c230-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:48:09 crc kubenswrapper[4872]: I1009 08:48:09.483298 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" event={"ID":"f0febe64-b028-4782-a3e8-31b07031c230","Type":"ContainerDied","Data":"2c89892086555bf8b0e90ce9438e21f2f4f1babbda7b78ce6dcaf349eda1278b"} Oct 09 08:48:09 crc kubenswrapper[4872]: I1009 08:48:09.483344 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c89892086555bf8b0e90ce9438e21f2f4f1babbda7b78ce6dcaf349eda1278b" Oct 09 08:48:09 crc kubenswrapper[4872]: I1009 08:48:09.483354 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-f24mq" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.046301 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv"] Oct 09 08:48:17 crc kubenswrapper[4872]: E1009 08:48:17.047452 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0febe64-b028-4782-a3e8-31b07031c230" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.047483 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0febe64-b028-4782-a3e8-31b07031c230" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.047739 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0febe64-b028-4782-a3e8-31b07031c230" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.048598 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.051886 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.052161 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.052396 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.052498 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.055625 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv"] Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.219980 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.220211 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.220273 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8jq5\" (UniqueName: \"kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.321877 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.322084 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.322138 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8jq5\" (UniqueName: \"kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.334931 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.336200 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.342469 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8jq5\" (UniqueName: \"kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.374870 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:48:17 crc kubenswrapper[4872]: I1009 08:48:17.911551 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv"] Oct 09 08:48:18 crc kubenswrapper[4872]: I1009 08:48:18.580862 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" event={"ID":"cc650d56-3744-41ca-90b1-707c8fc60bd9","Type":"ContainerStarted","Data":"05c08dd243a84c9eee460fe1414ed9c932efec21ccedcb002d2134335e075842"} Oct 09 08:48:19 crc kubenswrapper[4872]: I1009 08:48:19.462299 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:48:19 crc kubenswrapper[4872]: E1009 08:48:19.462701 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:48:19 crc kubenswrapper[4872]: I1009 08:48:19.589911 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" event={"ID":"cc650d56-3744-41ca-90b1-707c8fc60bd9","Type":"ContainerStarted","Data":"2d70f596dcf228c3248fe2ec5046a4be0732c3c8859e2d0cf4b4fd6e922f52cf"} Oct 09 08:48:19 crc kubenswrapper[4872]: I1009 08:48:19.612964 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" podStartSLOduration=2.110714837 podStartE2EDuration="2.612938089s" podCreationTimestamp="2025-10-09 08:48:17 +0000 UTC" firstStartedPulling="2025-10-09 08:48:17.918159489 +0000 UTC m=+1856.108688115" lastFinishedPulling="2025-10-09 08:48:18.420382741 +0000 UTC m=+1856.610911367" observedRunningTime="2025-10-09 08:48:19.60885031 +0000 UTC m=+1857.799378956" watchObservedRunningTime="2025-10-09 08:48:19.612938089 +0000 UTC m=+1857.803466715" Oct 09 08:48:34 crc kubenswrapper[4872]: I1009 08:48:34.461326 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:48:34 crc kubenswrapper[4872]: E1009 08:48:34.462286 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:48:37 crc kubenswrapper[4872]: I1009 08:48:37.049980 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-5db52"] Oct 09 08:48:37 crc kubenswrapper[4872]: I1009 08:48:37.056408 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-5db52"] Oct 09 08:48:38 crc kubenswrapper[4872]: I1009 08:48:38.475204 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e90359-2f2f-4f14-adf5-43f088b166a4" path="/var/lib/kubelet/pods/90e90359-2f2f-4f14-adf5-43f088b166a4/volumes" Oct 09 08:48:46 crc kubenswrapper[4872]: I1009 08:48:46.467313 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:48:46 crc kubenswrapper[4872]: I1009 08:48:46.812133 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde"} Oct 09 08:48:47 crc kubenswrapper[4872]: I1009 08:48:47.594436 4872 scope.go:117] "RemoveContainer" containerID="0b71167146630c71d28cbeb1eb3662716086ad25dfc4b1391d957e7be1a53cc6" Oct 09 08:48:47 crc kubenswrapper[4872]: I1009 08:48:47.659246 4872 scope.go:117] "RemoveContainer" containerID="73ce3c85dfcd9f73f5e0466e57db9103d94dfe90e36711c0083e2d6abdf2b25f" Oct 09 08:48:47 crc kubenswrapper[4872]: I1009 08:48:47.711957 4872 scope.go:117] "RemoveContainer" containerID="a8f8e67dd4074a3a0cfa9f4f8b289b0f734ae072d14085b0b850339887d460bc" Oct 09 08:48:59 crc kubenswrapper[4872]: I1009 08:48:59.948553 4872 generic.go:334] "Generic (PLEG): container finished" podID="cc650d56-3744-41ca-90b1-707c8fc60bd9" containerID="2d70f596dcf228c3248fe2ec5046a4be0732c3c8859e2d0cf4b4fd6e922f52cf" exitCode=0 Oct 09 08:48:59 crc kubenswrapper[4872]: I1009 08:48:59.949160 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" event={"ID":"cc650d56-3744-41ca-90b1-707c8fc60bd9","Type":"ContainerDied","Data":"2d70f596dcf228c3248fe2ec5046a4be0732c3c8859e2d0cf4b4fd6e922f52cf"} Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.337113 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.471348 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8jq5\" (UniqueName: \"kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5\") pod \"cc650d56-3744-41ca-90b1-707c8fc60bd9\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.471602 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key\") pod \"cc650d56-3744-41ca-90b1-707c8fc60bd9\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.471718 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory\") pod \"cc650d56-3744-41ca-90b1-707c8fc60bd9\" (UID: \"cc650d56-3744-41ca-90b1-707c8fc60bd9\") " Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.478263 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5" (OuterVolumeSpecName: "kube-api-access-p8jq5") pod "cc650d56-3744-41ca-90b1-707c8fc60bd9" (UID: "cc650d56-3744-41ca-90b1-707c8fc60bd9"). InnerVolumeSpecName "kube-api-access-p8jq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.503001 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory" (OuterVolumeSpecName: "inventory") pod "cc650d56-3744-41ca-90b1-707c8fc60bd9" (UID: "cc650d56-3744-41ca-90b1-707c8fc60bd9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.505932 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "cc650d56-3744-41ca-90b1-707c8fc60bd9" (UID: "cc650d56-3744-41ca-90b1-707c8fc60bd9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.574438 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8jq5\" (UniqueName: \"kubernetes.io/projected/cc650d56-3744-41ca-90b1-707c8fc60bd9-kube-api-access-p8jq5\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.574474 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.574483 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cc650d56-3744-41ca-90b1-707c8fc60bd9-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.970445 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" event={"ID":"cc650d56-3744-41ca-90b1-707c8fc60bd9","Type":"ContainerDied","Data":"05c08dd243a84c9eee460fe1414ed9c932efec21ccedcb002d2134335e075842"} Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.970492 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05c08dd243a84c9eee460fe1414ed9c932efec21ccedcb002d2134335e075842" Oct 09 08:49:01 crc kubenswrapper[4872]: I1009 08:49:01.970557 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.064245 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h292d"] Oct 09 08:49:02 crc kubenswrapper[4872]: E1009 08:49:02.064658 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc650d56-3744-41ca-90b1-707c8fc60bd9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.064674 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc650d56-3744-41ca-90b1-707c8fc60bd9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.064855 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc650d56-3744-41ca-90b1-707c8fc60bd9" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.065464 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.071040 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.071049 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.071558 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.071849 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.082046 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h292d"] Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.184996 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq97j\" (UniqueName: \"kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.185051 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.185158 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.286724 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq97j\" (UniqueName: \"kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.286786 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.286902 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.291498 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.291683 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.309693 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq97j\" (UniqueName: \"kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j\") pod \"ssh-known-hosts-edpm-deployment-h292d\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.394143 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.941451 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-h292d"] Oct 09 08:49:02 crc kubenswrapper[4872]: I1009 08:49:02.981191 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" event={"ID":"cefe7874-2185-4237-b9b6-aa8332849f11","Type":"ContainerStarted","Data":"e89927adf86a3a5ae5a40495fcdf058887de8ff09c992967ffdf900327e8e319"} Oct 09 08:49:03 crc kubenswrapper[4872]: I1009 08:49:03.993362 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" event={"ID":"cefe7874-2185-4237-b9b6-aa8332849f11","Type":"ContainerStarted","Data":"045f7236fe801923178b0d3c1e09bc5deb3d6a40c470cf75651ab165c6856a23"} Oct 09 08:49:04 crc kubenswrapper[4872]: I1009 08:49:04.018293 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" podStartSLOduration=1.593588289 podStartE2EDuration="2.01827372s" podCreationTimestamp="2025-10-09 08:49:02 +0000 UTC" firstStartedPulling="2025-10-09 08:49:02.938108396 +0000 UTC m=+1901.128637022" lastFinishedPulling="2025-10-09 08:49:03.362793797 +0000 UTC m=+1901.553322453" observedRunningTime="2025-10-09 08:49:04.011890935 +0000 UTC m=+1902.202419621" watchObservedRunningTime="2025-10-09 08:49:04.01827372 +0000 UTC m=+1902.208802346" Oct 09 08:49:11 crc kubenswrapper[4872]: I1009 08:49:11.052117 4872 generic.go:334] "Generic (PLEG): container finished" podID="cefe7874-2185-4237-b9b6-aa8332849f11" containerID="045f7236fe801923178b0d3c1e09bc5deb3d6a40c470cf75651ab165c6856a23" exitCode=0 Oct 09 08:49:11 crc kubenswrapper[4872]: I1009 08:49:11.052192 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" event={"ID":"cefe7874-2185-4237-b9b6-aa8332849f11","Type":"ContainerDied","Data":"045f7236fe801923178b0d3c1e09bc5deb3d6a40c470cf75651ab165c6856a23"} Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.450342 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.568369 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0\") pod \"cefe7874-2185-4237-b9b6-aa8332849f11\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.568488 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq97j\" (UniqueName: \"kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j\") pod \"cefe7874-2185-4237-b9b6-aa8332849f11\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.568686 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam\") pod \"cefe7874-2185-4237-b9b6-aa8332849f11\" (UID: \"cefe7874-2185-4237-b9b6-aa8332849f11\") " Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.576361 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j" (OuterVolumeSpecName: "kube-api-access-cq97j") pod "cefe7874-2185-4237-b9b6-aa8332849f11" (UID: "cefe7874-2185-4237-b9b6-aa8332849f11"). InnerVolumeSpecName "kube-api-access-cq97j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.600326 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cefe7874-2185-4237-b9b6-aa8332849f11" (UID: "cefe7874-2185-4237-b9b6-aa8332849f11"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.602385 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "cefe7874-2185-4237-b9b6-aa8332849f11" (UID: "cefe7874-2185-4237-b9b6-aa8332849f11"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.672182 4872 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-inventory-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.672233 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq97j\" (UniqueName: \"kubernetes.io/projected/cefe7874-2185-4237-b9b6-aa8332849f11-kube-api-access-cq97j\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:12 crc kubenswrapper[4872]: I1009 08:49:12.672246 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cefe7874-2185-4237-b9b6-aa8332849f11-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.070951 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" event={"ID":"cefe7874-2185-4237-b9b6-aa8332849f11","Type":"ContainerDied","Data":"e89927adf86a3a5ae5a40495fcdf058887de8ff09c992967ffdf900327e8e319"} Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.071024 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e89927adf86a3a5ae5a40495fcdf058887de8ff09c992967ffdf900327e8e319" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.071113 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-h292d" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.152984 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2"] Oct 09 08:49:13 crc kubenswrapper[4872]: E1009 08:49:13.153511 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cefe7874-2185-4237-b9b6-aa8332849f11" containerName="ssh-known-hosts-edpm-deployment" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.153535 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="cefe7874-2185-4237-b9b6-aa8332849f11" containerName="ssh-known-hosts-edpm-deployment" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.154632 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="cefe7874-2185-4237-b9b6-aa8332849f11" containerName="ssh-known-hosts-edpm-deployment" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.155434 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.158102 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.158129 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.159291 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.159336 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.163894 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2"] Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.282009 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.282252 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.282387 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qkm2\" (UniqueName: \"kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.384317 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.384464 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.384508 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qkm2\" (UniqueName: \"kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.389701 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.389736 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.401561 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qkm2\" (UniqueName: \"kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-w99z2\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.479418 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:13 crc kubenswrapper[4872]: I1009 08:49:13.981050 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2"] Oct 09 08:49:14 crc kubenswrapper[4872]: I1009 08:49:14.079790 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" event={"ID":"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d","Type":"ContainerStarted","Data":"b82dfc40fd5c13d0113bec800f76d038e241bedccbfb8601c4438324c4266051"} Oct 09 08:49:15 crc kubenswrapper[4872]: I1009 08:49:15.089837 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" event={"ID":"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d","Type":"ContainerStarted","Data":"eceb49a135409cfdff1da90d0a669dcc6718f7f01c34e5d581cb45488a979348"} Oct 09 08:49:15 crc kubenswrapper[4872]: I1009 08:49:15.116248 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" podStartSLOduration=1.4949697849999999 podStartE2EDuration="2.116221794s" podCreationTimestamp="2025-10-09 08:49:13 +0000 UTC" firstStartedPulling="2025-10-09 08:49:13.986244894 +0000 UTC m=+1912.176773520" lastFinishedPulling="2025-10-09 08:49:14.607496903 +0000 UTC m=+1912.798025529" observedRunningTime="2025-10-09 08:49:15.104387631 +0000 UTC m=+1913.294916277" watchObservedRunningTime="2025-10-09 08:49:15.116221794 +0000 UTC m=+1913.306750460" Oct 09 08:49:23 crc kubenswrapper[4872]: I1009 08:49:23.162615 4872 generic.go:334] "Generic (PLEG): container finished" podID="0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" containerID="eceb49a135409cfdff1da90d0a669dcc6718f7f01c34e5d581cb45488a979348" exitCode=0 Oct 09 08:49:23 crc kubenswrapper[4872]: I1009 08:49:23.162668 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" event={"ID":"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d","Type":"ContainerDied","Data":"eceb49a135409cfdff1da90d0a669dcc6718f7f01c34e5d581cb45488a979348"} Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.593546 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.703395 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qkm2\" (UniqueName: \"kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2\") pod \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.703453 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory\") pod \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.703554 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key\") pod \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\" (UID: \"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d\") " Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.708742 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2" (OuterVolumeSpecName: "kube-api-access-4qkm2") pod "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" (UID: "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d"). InnerVolumeSpecName "kube-api-access-4qkm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.729198 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" (UID: "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.730492 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory" (OuterVolumeSpecName: "inventory") pod "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" (UID: "0a5b2694-c0aa-44f4-aec1-4b3dad1a750d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.806118 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qkm2\" (UniqueName: \"kubernetes.io/projected/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-kube-api-access-4qkm2\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.806153 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:24 crc kubenswrapper[4872]: I1009 08:49:24.806163 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0a5b2694-c0aa-44f4-aec1-4b3dad1a750d-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.185083 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" event={"ID":"0a5b2694-c0aa-44f4-aec1-4b3dad1a750d","Type":"ContainerDied","Data":"b82dfc40fd5c13d0113bec800f76d038e241bedccbfb8601c4438324c4266051"} Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.185550 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b82dfc40fd5c13d0113bec800f76d038e241bedccbfb8601c4438324c4266051" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.185282 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-w99z2" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.265865 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll"] Oct 09 08:49:25 crc kubenswrapper[4872]: E1009 08:49:25.266253 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.266269 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.266458 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a5b2694-c0aa-44f4-aec1-4b3dad1a750d" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.267055 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.269341 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.270931 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.271144 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.273891 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.284008 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll"] Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.315521 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj4p5\" (UniqueName: \"kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.315971 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.316188 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.417390 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.417505 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.417652 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj4p5\" (UniqueName: \"kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.421535 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.422698 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.437049 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj4p5\" (UniqueName: \"kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:25 crc kubenswrapper[4872]: I1009 08:49:25.582746 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:49:26 crc kubenswrapper[4872]: I1009 08:49:26.105336 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll"] Oct 09 08:49:26 crc kubenswrapper[4872]: I1009 08:49:26.194297 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" event={"ID":"710632e7-e59c-4437-a9c5-171a80e989f6","Type":"ContainerStarted","Data":"1d07d5db2122453e4ac76752a5d3968e53b61e440baf248d1f5ef29470ec9db9"} Oct 09 08:49:27 crc kubenswrapper[4872]: I1009 08:49:27.229170 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" event={"ID":"710632e7-e59c-4437-a9c5-171a80e989f6","Type":"ContainerStarted","Data":"1bbdd06ad7352935e2a233911836cc91f9aa36743de44e8ce21e07cfcb53d531"} Oct 09 08:49:27 crc kubenswrapper[4872]: I1009 08:49:27.252142 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" podStartSLOduration=1.825213292 podStartE2EDuration="2.252126247s" podCreationTimestamp="2025-10-09 08:49:25 +0000 UTC" firstStartedPulling="2025-10-09 08:49:26.109510051 +0000 UTC m=+1924.300038677" lastFinishedPulling="2025-10-09 08:49:26.536423006 +0000 UTC m=+1924.726951632" observedRunningTime="2025-10-09 08:49:27.250187631 +0000 UTC m=+1925.440716267" watchObservedRunningTime="2025-10-09 08:49:27.252126247 +0000 UTC m=+1925.442654873" Oct 09 08:50:40 crc kubenswrapper[4872]: I1009 08:50:40.896255 4872 generic.go:334] "Generic (PLEG): container finished" podID="710632e7-e59c-4437-a9c5-171a80e989f6" containerID="1bbdd06ad7352935e2a233911836cc91f9aa36743de44e8ce21e07cfcb53d531" exitCode=0 Oct 09 08:50:40 crc kubenswrapper[4872]: I1009 08:50:40.896332 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" event={"ID":"710632e7-e59c-4437-a9c5-171a80e989f6","Type":"ContainerDied","Data":"1bbdd06ad7352935e2a233911836cc91f9aa36743de44e8ce21e07cfcb53d531"} Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.297327 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.299565 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key\") pod \"710632e7-e59c-4437-a9c5-171a80e989f6\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.299657 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory\") pod \"710632e7-e59c-4437-a9c5-171a80e989f6\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.331556 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory" (OuterVolumeSpecName: "inventory") pod "710632e7-e59c-4437-a9c5-171a80e989f6" (UID: "710632e7-e59c-4437-a9c5-171a80e989f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.335464 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "710632e7-e59c-4437-a9c5-171a80e989f6" (UID: "710632e7-e59c-4437-a9c5-171a80e989f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.400700 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj4p5\" (UniqueName: \"kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5\") pod \"710632e7-e59c-4437-a9c5-171a80e989f6\" (UID: \"710632e7-e59c-4437-a9c5-171a80e989f6\") " Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.401114 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.401134 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/710632e7-e59c-4437-a9c5-171a80e989f6-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.404390 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5" (OuterVolumeSpecName: "kube-api-access-rj4p5") pod "710632e7-e59c-4437-a9c5-171a80e989f6" (UID: "710632e7-e59c-4437-a9c5-171a80e989f6"). InnerVolumeSpecName "kube-api-access-rj4p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.503935 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj4p5\" (UniqueName: \"kubernetes.io/projected/710632e7-e59c-4437-a9c5-171a80e989f6-kube-api-access-rj4p5\") on node \"crc\" DevicePath \"\"" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.919778 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" event={"ID":"710632e7-e59c-4437-a9c5-171a80e989f6","Type":"ContainerDied","Data":"1d07d5db2122453e4ac76752a5d3968e53b61e440baf248d1f5ef29470ec9db9"} Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.919818 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d07d5db2122453e4ac76752a5d3968e53b61e440baf248d1f5ef29470ec9db9" Oct 09 08:50:42 crc kubenswrapper[4872]: I1009 08:50:42.920092 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.016250 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n"] Oct 09 08:50:43 crc kubenswrapper[4872]: E1009 08:50:43.018283 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="710632e7-e59c-4437-a9c5-171a80e989f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.018311 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="710632e7-e59c-4437-a9c5-171a80e989f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.018526 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="710632e7-e59c-4437-a9c5-171a80e989f6" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.019547 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.026842 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027037 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027163 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027272 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027385 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027503 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027607 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.027743 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.031828 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n"] Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.114524 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.114814 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.114946 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115106 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115154 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115205 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115262 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115299 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115333 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115369 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115420 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115456 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115483 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb968\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.115508 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217014 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217282 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217304 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217350 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217373 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217438 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217463 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217489 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217506 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217537 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217558 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217576 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb968\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.217622 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.222483 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.223332 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.223555 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.224871 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.225099 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.225611 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.225862 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.225910 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.228301 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.228374 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.229329 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.229583 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.232669 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.240259 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb968\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.353344 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.871712 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n"] Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.879027 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:50:43 crc kubenswrapper[4872]: I1009 08:50:43.928590 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" event={"ID":"abc08f25-2678-4241-9f86-d7c6700a7ae6","Type":"ContainerStarted","Data":"1e0e2e78a83d02381450b0f48db696c053e70a392263fc2eebe7bf5e2a90eab4"} Oct 09 08:50:44 crc kubenswrapper[4872]: I1009 08:50:44.940853 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" event={"ID":"abc08f25-2678-4241-9f86-d7c6700a7ae6","Type":"ContainerStarted","Data":"f8c97f16a780713504fa5d105bd8e476156097f62a59c28fb32f1156c18472d8"} Oct 09 08:50:44 crc kubenswrapper[4872]: I1009 08:50:44.966556 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" podStartSLOduration=2.503738973 podStartE2EDuration="2.966536489s" podCreationTimestamp="2025-10-09 08:50:42 +0000 UTC" firstStartedPulling="2025-10-09 08:50:43.87877219 +0000 UTC m=+2002.069300826" lastFinishedPulling="2025-10-09 08:50:44.341569706 +0000 UTC m=+2002.532098342" observedRunningTime="2025-10-09 08:50:44.958054421 +0000 UTC m=+2003.148583047" watchObservedRunningTime="2025-10-09 08:50:44.966536489 +0000 UTC m=+2003.157065115" Oct 09 08:51:08 crc kubenswrapper[4872]: I1009 08:51:08.157288 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:51:08 crc kubenswrapper[4872]: I1009 08:51:08.158133 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:51:21 crc kubenswrapper[4872]: I1009 08:51:21.256604 4872 generic.go:334] "Generic (PLEG): container finished" podID="abc08f25-2678-4241-9f86-d7c6700a7ae6" containerID="f8c97f16a780713504fa5d105bd8e476156097f62a59c28fb32f1156c18472d8" exitCode=0 Oct 09 08:51:21 crc kubenswrapper[4872]: I1009 08:51:21.256704 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" event={"ID":"abc08f25-2678-4241-9f86-d7c6700a7ae6","Type":"ContainerDied","Data":"f8c97f16a780713504fa5d105bd8e476156097f62a59c28fb32f1156c18472d8"} Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.663020 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782518 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782592 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782630 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782710 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782741 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782862 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782893 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vb968\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782929 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.782981 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.783005 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.783075 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.783171 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.783216 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.783247 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle\") pod \"abc08f25-2678-4241-9f86-d7c6700a7ae6\" (UID: \"abc08f25-2678-4241-9f86-d7c6700a7ae6\") " Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.791031 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794055 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794143 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794319 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794362 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794390 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968" (OuterVolumeSpecName: "kube-api-access-vb968") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "kube-api-access-vb968". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.794501 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.795053 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.795634 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.796365 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.799065 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.806915 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.822638 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.823076 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory" (OuterVolumeSpecName: "inventory") pod "abc08f25-2678-4241-9f86-d7c6700a7ae6" (UID: "abc08f25-2678-4241-9f86-d7c6700a7ae6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885569 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885615 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885627 4872 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885655 4872 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885667 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885678 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885686 4872 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885694 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885703 4872 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885712 4872 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885719 4872 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885727 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc08f25-2678-4241-9f86-d7c6700a7ae6-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885736 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:22 crc kubenswrapper[4872]: I1009 08:51:22.885746 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vb968\" (UniqueName: \"kubernetes.io/projected/abc08f25-2678-4241-9f86-d7c6700a7ae6-kube-api-access-vb968\") on node \"crc\" DevicePath \"\"" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.278892 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" event={"ID":"abc08f25-2678-4241-9f86-d7c6700a7ae6","Type":"ContainerDied","Data":"1e0e2e78a83d02381450b0f48db696c053e70a392263fc2eebe7bf5e2a90eab4"} Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.278935 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e0e2e78a83d02381450b0f48db696c053e70a392263fc2eebe7bf5e2a90eab4" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.279009 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.394856 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw"] Oct 09 08:51:23 crc kubenswrapper[4872]: E1009 08:51:23.395320 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc08f25-2678-4241-9f86-d7c6700a7ae6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.395343 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc08f25-2678-4241-9f86-d7c6700a7ae6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.395613 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc08f25-2678-4241-9f86-d7c6700a7ae6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.396442 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.400680 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.401006 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.401470 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.401872 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.401970 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.413514 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw"] Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.497531 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.497655 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.497781 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.497830 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqbdr\" (UniqueName: \"kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.497875 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.599392 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.599465 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.599537 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.599575 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqbdr\" (UniqueName: \"kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.599605 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.600932 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.604449 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.605459 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.607952 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.617322 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqbdr\" (UniqueName: \"kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-wwvpw\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:23 crc kubenswrapper[4872]: I1009 08:51:23.711824 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:51:24 crc kubenswrapper[4872]: I1009 08:51:24.225274 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw"] Oct 09 08:51:24 crc kubenswrapper[4872]: I1009 08:51:24.289855 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" event={"ID":"d9ac9cf3-10a8-45ef-b729-4822000bdc63","Type":"ContainerStarted","Data":"1dfed4e996fd985a984c2762706a70649b57dcd7c76a02d40bd0f51a9350cdc3"} Oct 09 08:51:25 crc kubenswrapper[4872]: I1009 08:51:25.299166 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" event={"ID":"d9ac9cf3-10a8-45ef-b729-4822000bdc63","Type":"ContainerStarted","Data":"5b4977df03b7b50ee413db2c6211343ecd527bcd93dcdd317e59026c7cede7dd"} Oct 09 08:51:25 crc kubenswrapper[4872]: I1009 08:51:25.314767 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" podStartSLOduration=1.8008677149999999 podStartE2EDuration="2.314747189s" podCreationTimestamp="2025-10-09 08:51:23 +0000 UTC" firstStartedPulling="2025-10-09 08:51:24.230126543 +0000 UTC m=+2042.420655179" lastFinishedPulling="2025-10-09 08:51:24.744006027 +0000 UTC m=+2042.934534653" observedRunningTime="2025-10-09 08:51:25.311761572 +0000 UTC m=+2043.502290218" watchObservedRunningTime="2025-10-09 08:51:25.314747189 +0000 UTC m=+2043.505275815" Oct 09 08:51:38 crc kubenswrapper[4872]: I1009 08:51:38.156951 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:51:38 crc kubenswrapper[4872]: I1009 08:51:38.157552 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.233982 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.237061 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.243332 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.352589 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.352903 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.353089 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phf5l\" (UniqueName: \"kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.455232 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.455302 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.455333 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phf5l\" (UniqueName: \"kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.455970 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.455990 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.483795 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phf5l\" (UniqueName: \"kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l\") pod \"certified-operators-7h2tl\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:57 crc kubenswrapper[4872]: I1009 08:51:57.559941 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:51:58 crc kubenswrapper[4872]: I1009 08:51:58.137562 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:51:58 crc kubenswrapper[4872]: I1009 08:51:58.625285 4872 generic.go:334] "Generic (PLEG): container finished" podID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerID="eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b" exitCode=0 Oct 09 08:51:58 crc kubenswrapper[4872]: I1009 08:51:58.625445 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerDied","Data":"eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b"} Oct 09 08:51:58 crc kubenswrapper[4872]: I1009 08:51:58.625559 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerStarted","Data":"d7a052759e54e1d37bd8b1a036633eabbee5b4c4dc8b3db2df0022d41019c4bc"} Oct 09 08:51:59 crc kubenswrapper[4872]: I1009 08:51:59.637401 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerStarted","Data":"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f"} Oct 09 08:52:01 crc kubenswrapper[4872]: I1009 08:52:01.664921 4872 generic.go:334] "Generic (PLEG): container finished" podID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerID="849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f" exitCode=0 Oct 09 08:52:01 crc kubenswrapper[4872]: I1009 08:52:01.665042 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerDied","Data":"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f"} Oct 09 08:52:02 crc kubenswrapper[4872]: I1009 08:52:02.676751 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerStarted","Data":"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550"} Oct 09 08:52:02 crc kubenswrapper[4872]: I1009 08:52:02.695355 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7h2tl" podStartSLOduration=2.061480591 podStartE2EDuration="5.695337824s" podCreationTimestamp="2025-10-09 08:51:57 +0000 UTC" firstStartedPulling="2025-10-09 08:51:58.627127873 +0000 UTC m=+2076.817656499" lastFinishedPulling="2025-10-09 08:52:02.260985106 +0000 UTC m=+2080.451513732" observedRunningTime="2025-10-09 08:52:02.694716946 +0000 UTC m=+2080.885245592" watchObservedRunningTime="2025-10-09 08:52:02.695337824 +0000 UTC m=+2080.885866450" Oct 09 08:52:07 crc kubenswrapper[4872]: I1009 08:52:07.560281 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:07 crc kubenswrapper[4872]: I1009 08:52:07.560903 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:07 crc kubenswrapper[4872]: I1009 08:52:07.609729 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:07 crc kubenswrapper[4872]: I1009 08:52:07.765249 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:07 crc kubenswrapper[4872]: I1009 08:52:07.845404 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.156713 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.156800 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.156892 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.157925 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.158006 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde" gracePeriod=600 Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.738080 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde" exitCode=0 Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.738149 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde"} Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.738464 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a"} Oct 09 08:52:08 crc kubenswrapper[4872]: I1009 08:52:08.738500 4872 scope.go:117] "RemoveContainer" containerID="b486797b5216e75992dc9313b865e28442b8b5da07fcd305cea6c2d958f0c549" Oct 09 08:52:09 crc kubenswrapper[4872]: I1009 08:52:09.756171 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7h2tl" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="registry-server" containerID="cri-o://8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550" gracePeriod=2 Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.200128 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.307523 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities\") pod \"09942565-bd30-4cbc-b06e-5e70d1650c65\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.307723 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content\") pod \"09942565-bd30-4cbc-b06e-5e70d1650c65\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.307754 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phf5l\" (UniqueName: \"kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l\") pod \"09942565-bd30-4cbc-b06e-5e70d1650c65\" (UID: \"09942565-bd30-4cbc-b06e-5e70d1650c65\") " Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.308743 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities" (OuterVolumeSpecName: "utilities") pod "09942565-bd30-4cbc-b06e-5e70d1650c65" (UID: "09942565-bd30-4cbc-b06e-5e70d1650c65"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.313990 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l" (OuterVolumeSpecName: "kube-api-access-phf5l") pod "09942565-bd30-4cbc-b06e-5e70d1650c65" (UID: "09942565-bd30-4cbc-b06e-5e70d1650c65"). InnerVolumeSpecName "kube-api-access-phf5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.364914 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "09942565-bd30-4cbc-b06e-5e70d1650c65" (UID: "09942565-bd30-4cbc-b06e-5e70d1650c65"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.410379 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.410425 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/09942565-bd30-4cbc-b06e-5e70d1650c65-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.410441 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phf5l\" (UniqueName: \"kubernetes.io/projected/09942565-bd30-4cbc-b06e-5e70d1650c65-kube-api-access-phf5l\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.768351 4872 generic.go:334] "Generic (PLEG): container finished" podID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerID="8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550" exitCode=0 Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.768423 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7h2tl" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.768408 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerDied","Data":"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550"} Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.769210 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7h2tl" event={"ID":"09942565-bd30-4cbc-b06e-5e70d1650c65","Type":"ContainerDied","Data":"d7a052759e54e1d37bd8b1a036633eabbee5b4c4dc8b3db2df0022d41019c4bc"} Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.769235 4872 scope.go:117] "RemoveContainer" containerID="8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.801250 4872 scope.go:117] "RemoveContainer" containerID="849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.805101 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.815212 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7h2tl"] Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.819310 4872 scope.go:117] "RemoveContainer" containerID="eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.861768 4872 scope.go:117] "RemoveContainer" containerID="8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550" Oct 09 08:52:10 crc kubenswrapper[4872]: E1009 08:52:10.862217 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550\": container with ID starting with 8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550 not found: ID does not exist" containerID="8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.862257 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550"} err="failed to get container status \"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550\": rpc error: code = NotFound desc = could not find container \"8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550\": container with ID starting with 8424ef192a941a193ea1cd87cf346e906705ef3ccc8452bb101d33ab3d2f6550 not found: ID does not exist" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.862279 4872 scope.go:117] "RemoveContainer" containerID="849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f" Oct 09 08:52:10 crc kubenswrapper[4872]: E1009 08:52:10.862579 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f\": container with ID starting with 849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f not found: ID does not exist" containerID="849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.862618 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f"} err="failed to get container status \"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f\": rpc error: code = NotFound desc = could not find container \"849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f\": container with ID starting with 849917f2ea346c683c57b94bbc9b5f7bd1e3fa8f24b5eb91e24722f84cae2e3f not found: ID does not exist" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.862635 4872 scope.go:117] "RemoveContainer" containerID="eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b" Oct 09 08:52:10 crc kubenswrapper[4872]: E1009 08:52:10.863072 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b\": container with ID starting with eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b not found: ID does not exist" containerID="eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b" Oct 09 08:52:10 crc kubenswrapper[4872]: I1009 08:52:10.863130 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b"} err="failed to get container status \"eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b\": rpc error: code = NotFound desc = could not find container \"eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b\": container with ID starting with eae17c78433c1f94a7752bfe66a41250f4b0bef6546fac35385aa0c0a63a3e7b not found: ID does not exist" Oct 09 08:52:12 crc kubenswrapper[4872]: I1009 08:52:12.472326 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" path="/var/lib/kubelet/pods/09942565-bd30-4cbc-b06e-5e70d1650c65/volumes" Oct 09 08:52:30 crc kubenswrapper[4872]: I1009 08:52:30.959081 4872 generic.go:334] "Generic (PLEG): container finished" podID="d9ac9cf3-10a8-45ef-b729-4822000bdc63" containerID="5b4977df03b7b50ee413db2c6211343ecd527bcd93dcdd317e59026c7cede7dd" exitCode=0 Oct 09 08:52:30 crc kubenswrapper[4872]: I1009 08:52:30.959307 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" event={"ID":"d9ac9cf3-10a8-45ef-b729-4822000bdc63","Type":"ContainerDied","Data":"5b4977df03b7b50ee413db2c6211343ecd527bcd93dcdd317e59026c7cede7dd"} Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.366037 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.454423 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory\") pod \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.454806 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle\") pod \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.454901 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key\") pod \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.454958 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0\") pod \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.455018 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqbdr\" (UniqueName: \"kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr\") pod \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\" (UID: \"d9ac9cf3-10a8-45ef-b729-4822000bdc63\") " Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.461662 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d9ac9cf3-10a8-45ef-b729-4822000bdc63" (UID: "d9ac9cf3-10a8-45ef-b729-4822000bdc63"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.464875 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr" (OuterVolumeSpecName: "kube-api-access-nqbdr") pod "d9ac9cf3-10a8-45ef-b729-4822000bdc63" (UID: "d9ac9cf3-10a8-45ef-b729-4822000bdc63"). InnerVolumeSpecName "kube-api-access-nqbdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.480673 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "d9ac9cf3-10a8-45ef-b729-4822000bdc63" (UID: "d9ac9cf3-10a8-45ef-b729-4822000bdc63"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.482653 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory" (OuterVolumeSpecName: "inventory") pod "d9ac9cf3-10a8-45ef-b729-4822000bdc63" (UID: "d9ac9cf3-10a8-45ef-b729-4822000bdc63"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.492694 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d9ac9cf3-10a8-45ef-b729-4822000bdc63" (UID: "d9ac9cf3-10a8-45ef-b729-4822000bdc63"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.557632 4872 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.557686 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.557700 4872 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/d9ac9cf3-10a8-45ef-b729-4822000bdc63-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.557710 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqbdr\" (UniqueName: \"kubernetes.io/projected/d9ac9cf3-10a8-45ef-b729-4822000bdc63-kube-api-access-nqbdr\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.557721 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d9ac9cf3-10a8-45ef-b729-4822000bdc63-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.977335 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" event={"ID":"d9ac9cf3-10a8-45ef-b729-4822000bdc63","Type":"ContainerDied","Data":"1dfed4e996fd985a984c2762706a70649b57dcd7c76a02d40bd0f51a9350cdc3"} Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.977385 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1dfed4e996fd985a984c2762706a70649b57dcd7c76a02d40bd0f51a9350cdc3" Oct 09 08:52:32 crc kubenswrapper[4872]: I1009 08:52:32.977410 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-wwvpw" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.062380 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl"] Oct 09 08:52:33 crc kubenswrapper[4872]: E1009 08:52:33.062794 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9ac9cf3-10a8-45ef-b729-4822000bdc63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.062812 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9ac9cf3-10a8-45ef-b729-4822000bdc63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 08:52:33 crc kubenswrapper[4872]: E1009 08:52:33.062832 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="extract-content" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.062838 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="extract-content" Oct 09 08:52:33 crc kubenswrapper[4872]: E1009 08:52:33.062869 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="extract-utilities" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.062874 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="extract-utilities" Oct 09 08:52:33 crc kubenswrapper[4872]: E1009 08:52:33.062885 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="registry-server" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.062891 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="registry-server" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.063083 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9ac9cf3-10a8-45ef-b729-4822000bdc63" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.063110 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="09942565-bd30-4cbc-b06e-5e70d1650c65" containerName="registry-server" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.063752 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.066356 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.066481 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.066523 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.066669 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.066672 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.068898 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.074097 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl"] Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169059 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169351 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169471 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169535 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169835 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.169878 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgrbq\" (UniqueName: \"kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271173 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271256 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271283 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271347 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271366 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgrbq\" (UniqueName: \"kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.271400 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.275207 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.275268 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.291699 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.291775 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.291889 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.298597 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgrbq\" (UniqueName: \"kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.380560 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.881517 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl"] Oct 09 08:52:33 crc kubenswrapper[4872]: I1009 08:52:33.986755 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" event={"ID":"92e98033-47ad-4e23-a3c4-d8c886be6d0c","Type":"ContainerStarted","Data":"e3457b23f93cedcc6e1025ddb72a5a9b01cc9e51dab4776039fa689d531acd6d"} Oct 09 08:52:34 crc kubenswrapper[4872]: I1009 08:52:34.999044 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" event={"ID":"92e98033-47ad-4e23-a3c4-d8c886be6d0c","Type":"ContainerStarted","Data":"867fd89ac3cf28a206092ac2613eb7466e30063dc18f87a94e5b608a49ea6718"} Oct 09 08:52:35 crc kubenswrapper[4872]: I1009 08:52:35.022625 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" podStartSLOduration=1.219334252 podStartE2EDuration="2.02260211s" podCreationTimestamp="2025-10-09 08:52:33 +0000 UTC" firstStartedPulling="2025-10-09 08:52:33.881826856 +0000 UTC m=+2112.072355482" lastFinishedPulling="2025-10-09 08:52:34.685094714 +0000 UTC m=+2112.875623340" observedRunningTime="2025-10-09 08:52:35.018570632 +0000 UTC m=+2113.209099288" watchObservedRunningTime="2025-10-09 08:52:35.02260211 +0000 UTC m=+2113.213130746" Oct 09 08:53:19 crc kubenswrapper[4872]: I1009 08:53:19.393695 4872 generic.go:334] "Generic (PLEG): container finished" podID="92e98033-47ad-4e23-a3c4-d8c886be6d0c" containerID="867fd89ac3cf28a206092ac2613eb7466e30063dc18f87a94e5b608a49ea6718" exitCode=0 Oct 09 08:53:19 crc kubenswrapper[4872]: I1009 08:53:19.393766 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" event={"ID":"92e98033-47ad-4e23-a3c4-d8c886be6d0c","Type":"ContainerDied","Data":"867fd89ac3cf28a206092ac2613eb7466e30063dc18f87a94e5b608a49ea6718"} Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.801620 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.903863 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.904412 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgrbq\" (UniqueName: \"kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.904629 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.904760 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.904927 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.905072 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key\") pod \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\" (UID: \"92e98033-47ad-4e23-a3c4-d8c886be6d0c\") " Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.911178 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.911258 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq" (OuterVolumeSpecName: "kube-api-access-pgrbq") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "kube-api-access-pgrbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.934700 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.936394 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.940090 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:53:20 crc kubenswrapper[4872]: I1009 08:53:20.952436 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory" (OuterVolumeSpecName: "inventory") pod "92e98033-47ad-4e23-a3c4-d8c886be6d0c" (UID: "92e98033-47ad-4e23-a3c4-d8c886be6d0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007581 4872 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007619 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007630 4872 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007693 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgrbq\" (UniqueName: \"kubernetes.io/projected/92e98033-47ad-4e23-a3c4-d8c886be6d0c-kube-api-access-pgrbq\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007702 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.007710 4872 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92e98033-47ad-4e23-a3c4-d8c886be6d0c-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.415440 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" event={"ID":"92e98033-47ad-4e23-a3c4-d8c886be6d0c","Type":"ContainerDied","Data":"e3457b23f93cedcc6e1025ddb72a5a9b01cc9e51dab4776039fa689d531acd6d"} Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.415487 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3457b23f93cedcc6e1025ddb72a5a9b01cc9e51dab4776039fa689d531acd6d" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.415747 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.546298 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq"] Oct 09 08:53:21 crc kubenswrapper[4872]: E1009 08:53:21.547070 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e98033-47ad-4e23-a3c4-d8c886be6d0c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.547093 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e98033-47ad-4e23-a3c4-d8c886be6d0c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.547293 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e98033-47ad-4e23-a3c4-d8c886be6d0c" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.548027 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.550751 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.550897 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.551063 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.551230 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.551327 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.556886 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq"] Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.619953 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.620000 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfkjt\" (UniqueName: \"kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.620054 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.620107 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.620144 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.721992 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.722336 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfkjt\" (UniqueName: \"kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.722476 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.722620 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.722808 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.727100 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.727398 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.727540 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.727861 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.741277 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfkjt\" (UniqueName: \"kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-dvntq\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:21 crc kubenswrapper[4872]: I1009 08:53:21.872473 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.434378 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq"] Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.576048 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.578615 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.586329 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.639478 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.639816 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5458s\" (UniqueName: \"kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.639982 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.741770 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.742127 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5458s\" (UniqueName: \"kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.742263 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.742407 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.742654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.765558 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5458s\" (UniqueName: \"kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s\") pod \"community-operators-ztpm6\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:22 crc kubenswrapper[4872]: I1009 08:53:22.905089 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:23 crc kubenswrapper[4872]: I1009 08:53:23.083612 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:53:23 crc kubenswrapper[4872]: I1009 08:53:23.433632 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" event={"ID":"80ca0a43-215b-42e4-8278-c7ded62e8080","Type":"ContainerStarted","Data":"e481328e7ea274287cc6995242b09c994010b44f9c5700380a80375bf3987976"} Oct 09 08:53:23 crc kubenswrapper[4872]: I1009 08:53:23.433695 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" event={"ID":"80ca0a43-215b-42e4-8278-c7ded62e8080","Type":"ContainerStarted","Data":"7b7ad4c5e651ec1e157079a87ca009bc1e9f0c0c712508af59307e4eb60ad3cf"} Oct 09 08:53:23 crc kubenswrapper[4872]: I1009 08:53:23.454204 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" podStartSLOduration=1.81494504 podStartE2EDuration="2.454182827s" podCreationTimestamp="2025-10-09 08:53:21 +0000 UTC" firstStartedPulling="2025-10-09 08:53:22.441974451 +0000 UTC m=+2160.632503087" lastFinishedPulling="2025-10-09 08:53:23.081212238 +0000 UTC m=+2161.271740874" observedRunningTime="2025-10-09 08:53:23.451571701 +0000 UTC m=+2161.642100327" watchObservedRunningTime="2025-10-09 08:53:23.454182827 +0000 UTC m=+2161.644711463" Oct 09 08:53:23 crc kubenswrapper[4872]: I1009 08:53:23.496449 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:23 crc kubenswrapper[4872]: W1009 08:53:23.504714 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod86c4b5aa_4517_4ca3_bfbd_b739bd1caa90.slice/crio-e875dbb23dedfb53e4e37425ce6a90cd75aec573e70ef1d6bd42a99ae1b50978 WatchSource:0}: Error finding container e875dbb23dedfb53e4e37425ce6a90cd75aec573e70ef1d6bd42a99ae1b50978: Status 404 returned error can't find the container with id e875dbb23dedfb53e4e37425ce6a90cd75aec573e70ef1d6bd42a99ae1b50978 Oct 09 08:53:24 crc kubenswrapper[4872]: I1009 08:53:24.444980 4872 generic.go:334] "Generic (PLEG): container finished" podID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerID="ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7" exitCode=0 Oct 09 08:53:24 crc kubenswrapper[4872]: I1009 08:53:24.445048 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerDied","Data":"ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7"} Oct 09 08:53:24 crc kubenswrapper[4872]: I1009 08:53:24.445372 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerStarted","Data":"e875dbb23dedfb53e4e37425ce6a90cd75aec573e70ef1d6bd42a99ae1b50978"} Oct 09 08:53:25 crc kubenswrapper[4872]: I1009 08:53:25.456229 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerStarted","Data":"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4"} Oct 09 08:53:26 crc kubenswrapper[4872]: I1009 08:53:26.467334 4872 generic.go:334] "Generic (PLEG): container finished" podID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerID="bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4" exitCode=0 Oct 09 08:53:26 crc kubenswrapper[4872]: I1009 08:53:26.478952 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerDied","Data":"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4"} Oct 09 08:53:27 crc kubenswrapper[4872]: I1009 08:53:27.483097 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerStarted","Data":"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779"} Oct 09 08:53:32 crc kubenswrapper[4872]: I1009 08:53:32.905585 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:32 crc kubenswrapper[4872]: I1009 08:53:32.905917 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:32 crc kubenswrapper[4872]: I1009 08:53:32.963211 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:32 crc kubenswrapper[4872]: I1009 08:53:32.984551 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ztpm6" podStartSLOduration=8.487353868 podStartE2EDuration="10.984532247s" podCreationTimestamp="2025-10-09 08:53:22 +0000 UTC" firstStartedPulling="2025-10-09 08:53:24.446661608 +0000 UTC m=+2162.637190234" lastFinishedPulling="2025-10-09 08:53:26.943839987 +0000 UTC m=+2165.134368613" observedRunningTime="2025-10-09 08:53:27.501858328 +0000 UTC m=+2165.692386964" watchObservedRunningTime="2025-10-09 08:53:32.984532247 +0000 UTC m=+2171.175060883" Oct 09 08:53:33 crc kubenswrapper[4872]: I1009 08:53:33.584870 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:34 crc kubenswrapper[4872]: I1009 08:53:34.348303 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:35 crc kubenswrapper[4872]: I1009 08:53:35.558191 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ztpm6" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="registry-server" containerID="cri-o://0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779" gracePeriod=2 Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.093541 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.224234 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content\") pod \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.224625 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5458s\" (UniqueName: \"kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s\") pod \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.224768 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities\") pod \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\" (UID: \"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90\") " Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.225715 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities" (OuterVolumeSpecName: "utilities") pod "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" (UID: "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.232607 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s" (OuterVolumeSpecName: "kube-api-access-5458s") pod "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" (UID: "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90"). InnerVolumeSpecName "kube-api-access-5458s". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.278333 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" (UID: "86c4b5aa-4517-4ca3-bfbd-b739bd1caa90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.326994 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.327029 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5458s\" (UniqueName: \"kubernetes.io/projected/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-kube-api-access-5458s\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.327041 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.568552 4872 generic.go:334] "Generic (PLEG): container finished" podID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerID="0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779" exitCode=0 Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.568617 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ztpm6" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.568618 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerDied","Data":"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779"} Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.568779 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ztpm6" event={"ID":"86c4b5aa-4517-4ca3-bfbd-b739bd1caa90","Type":"ContainerDied","Data":"e875dbb23dedfb53e4e37425ce6a90cd75aec573e70ef1d6bd42a99ae1b50978"} Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.568821 4872 scope.go:117] "RemoveContainer" containerID="0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.600786 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.603252 4872 scope.go:117] "RemoveContainer" containerID="bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.614732 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ztpm6"] Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.623772 4872 scope.go:117] "RemoveContainer" containerID="ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.679447 4872 scope.go:117] "RemoveContainer" containerID="0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779" Oct 09 08:53:36 crc kubenswrapper[4872]: E1009 08:53:36.679925 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779\": container with ID starting with 0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779 not found: ID does not exist" containerID="0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.679989 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779"} err="failed to get container status \"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779\": rpc error: code = NotFound desc = could not find container \"0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779\": container with ID starting with 0465a13d63031f745302184e96951acd3931e5267f790fcef9668979bf2a2779 not found: ID does not exist" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.680033 4872 scope.go:117] "RemoveContainer" containerID="bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4" Oct 09 08:53:36 crc kubenswrapper[4872]: E1009 08:53:36.680669 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4\": container with ID starting with bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4 not found: ID does not exist" containerID="bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.680714 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4"} err="failed to get container status \"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4\": rpc error: code = NotFound desc = could not find container \"bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4\": container with ID starting with bc73e0cf429a2725870b131ba6bbad882c46a1f6585070ca655ec72adc0e51e4 not found: ID does not exist" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.680739 4872 scope.go:117] "RemoveContainer" containerID="ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7" Oct 09 08:53:36 crc kubenswrapper[4872]: E1009 08:53:36.681272 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7\": container with ID starting with ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7 not found: ID does not exist" containerID="ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7" Oct 09 08:53:36 crc kubenswrapper[4872]: I1009 08:53:36.681320 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7"} err="failed to get container status \"ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7\": rpc error: code = NotFound desc = could not find container \"ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7\": container with ID starting with ca24347eb13ee4c27e020e263318778071d5cf4cff39d384f40bc8732af776a7 not found: ID does not exist" Oct 09 08:53:38 crc kubenswrapper[4872]: I1009 08:53:38.472618 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" path="/var/lib/kubelet/pods/86c4b5aa-4517-4ca3-bfbd-b739bd1caa90/volumes" Oct 09 08:54:08 crc kubenswrapper[4872]: I1009 08:54:08.157410 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:54:08 crc kubenswrapper[4872]: I1009 08:54:08.158109 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:54:38 crc kubenswrapper[4872]: I1009 08:54:38.157126 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:54:38 crc kubenswrapper[4872]: I1009 08:54:38.157826 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.157580 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.158155 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.158202 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.158749 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.158816 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" gracePeriod=600 Oct 09 08:55:08 crc kubenswrapper[4872]: E1009 08:55:08.283487 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.425609 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" exitCode=0 Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.425676 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a"} Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.425751 4872 scope.go:117] "RemoveContainer" containerID="cf3df92f1d453cf8b46e8dab49cbb0deb2b632f7e7a692c9da5859da0c12bfde" Oct 09 08:55:08 crc kubenswrapper[4872]: I1009 08:55:08.426349 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:55:08 crc kubenswrapper[4872]: E1009 08:55:08.426622 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:55:20 crc kubenswrapper[4872]: I1009 08:55:20.462364 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:55:20 crc kubenswrapper[4872]: E1009 08:55:20.464566 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:55:35 crc kubenswrapper[4872]: I1009 08:55:35.463310 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:55:35 crc kubenswrapper[4872]: E1009 08:55:35.464244 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:55:50 crc kubenswrapper[4872]: I1009 08:55:50.462012 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:55:50 crc kubenswrapper[4872]: E1009 08:55:50.463135 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.586909 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:55:52 crc kubenswrapper[4872]: E1009 08:55:52.587742 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="registry-server" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.587760 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="registry-server" Oct 09 08:55:52 crc kubenswrapper[4872]: E1009 08:55:52.587781 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="extract-utilities" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.587789 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="extract-utilities" Oct 09 08:55:52 crc kubenswrapper[4872]: E1009 08:55:52.587829 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="extract-content" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.587838 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="extract-content" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.588061 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="86c4b5aa-4517-4ca3-bfbd-b739bd1caa90" containerName="registry-server" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.589899 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.599302 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.750342 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.750545 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.750589 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49nzj\" (UniqueName: \"kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.859033 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.859108 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.859129 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49nzj\" (UniqueName: \"kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.859721 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.859733 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.889184 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49nzj\" (UniqueName: \"kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj\") pod \"redhat-marketplace-k88ld\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:52 crc kubenswrapper[4872]: I1009 08:55:52.922205 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:55:53 crc kubenswrapper[4872]: I1009 08:55:53.400735 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:55:53 crc kubenswrapper[4872]: I1009 08:55:53.836003 4872 generic.go:334] "Generic (PLEG): container finished" podID="77415b07-3938-4072-805b-64fff6789e72" containerID="0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff" exitCode=0 Oct 09 08:55:53 crc kubenswrapper[4872]: I1009 08:55:53.836068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerDied","Data":"0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff"} Oct 09 08:55:53 crc kubenswrapper[4872]: I1009 08:55:53.836133 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerStarted","Data":"02b5c6bdb19e72662d7413df71224968aa26b2cfa02f449de8e8bacc6fd79702"} Oct 09 08:55:53 crc kubenswrapper[4872]: I1009 08:55:53.837776 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 08:55:54 crc kubenswrapper[4872]: I1009 08:55:54.849340 4872 generic.go:334] "Generic (PLEG): container finished" podID="77415b07-3938-4072-805b-64fff6789e72" containerID="66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862" exitCode=0 Oct 09 08:55:54 crc kubenswrapper[4872]: I1009 08:55:54.849456 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerDied","Data":"66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862"} Oct 09 08:55:55 crc kubenswrapper[4872]: I1009 08:55:55.861766 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerStarted","Data":"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7"} Oct 09 08:56:02 crc kubenswrapper[4872]: I1009 08:56:02.922415 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:02 crc kubenswrapper[4872]: I1009 08:56:02.922889 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:02 crc kubenswrapper[4872]: I1009 08:56:02.979838 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:03 crc kubenswrapper[4872]: I1009 08:56:03.006775 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k88ld" podStartSLOduration=9.551425071 podStartE2EDuration="11.006753211s" podCreationTimestamp="2025-10-09 08:55:52 +0000 UTC" firstStartedPulling="2025-10-09 08:55:53.837543375 +0000 UTC m=+2312.028072001" lastFinishedPulling="2025-10-09 08:55:55.292871515 +0000 UTC m=+2313.483400141" observedRunningTime="2025-10-09 08:55:55.884241295 +0000 UTC m=+2314.074769931" watchObservedRunningTime="2025-10-09 08:56:03.006753211 +0000 UTC m=+2321.197281847" Oct 09 08:56:03 crc kubenswrapper[4872]: I1009 08:56:03.980687 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:04 crc kubenswrapper[4872]: I1009 08:56:04.032581 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:56:04 crc kubenswrapper[4872]: I1009 08:56:04.462245 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:56:04 crc kubenswrapper[4872]: E1009 08:56:04.462698 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:56:05 crc kubenswrapper[4872]: I1009 08:56:05.955438 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k88ld" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="registry-server" containerID="cri-o://d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7" gracePeriod=2 Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.401068 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.527052 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content\") pod \"77415b07-3938-4072-805b-64fff6789e72\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.527236 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities\") pod \"77415b07-3938-4072-805b-64fff6789e72\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.527459 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49nzj\" (UniqueName: \"kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj\") pod \"77415b07-3938-4072-805b-64fff6789e72\" (UID: \"77415b07-3938-4072-805b-64fff6789e72\") " Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.528881 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities" (OuterVolumeSpecName: "utilities") pod "77415b07-3938-4072-805b-64fff6789e72" (UID: "77415b07-3938-4072-805b-64fff6789e72"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.535064 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj" (OuterVolumeSpecName: "kube-api-access-49nzj") pod "77415b07-3938-4072-805b-64fff6789e72" (UID: "77415b07-3938-4072-805b-64fff6789e72"). InnerVolumeSpecName "kube-api-access-49nzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.548772 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "77415b07-3938-4072-805b-64fff6789e72" (UID: "77415b07-3938-4072-805b-64fff6789e72"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.631388 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.631423 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/77415b07-3938-4072-805b-64fff6789e72-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.631438 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49nzj\" (UniqueName: \"kubernetes.io/projected/77415b07-3938-4072-805b-64fff6789e72-kube-api-access-49nzj\") on node \"crc\" DevicePath \"\"" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.967823 4872 generic.go:334] "Generic (PLEG): container finished" podID="77415b07-3938-4072-805b-64fff6789e72" containerID="d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7" exitCode=0 Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.967895 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerDied","Data":"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7"} Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.967928 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k88ld" event={"ID":"77415b07-3938-4072-805b-64fff6789e72","Type":"ContainerDied","Data":"02b5c6bdb19e72662d7413df71224968aa26b2cfa02f449de8e8bacc6fd79702"} Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.967932 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k88ld" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.967969 4872 scope.go:117] "RemoveContainer" containerID="d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7" Oct 09 08:56:06 crc kubenswrapper[4872]: I1009 08:56:06.990948 4872 scope.go:117] "RemoveContainer" containerID="66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.008152 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.017273 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k88ld"] Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.026393 4872 scope.go:117] "RemoveContainer" containerID="0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.072264 4872 scope.go:117] "RemoveContainer" containerID="d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7" Oct 09 08:56:07 crc kubenswrapper[4872]: E1009 08:56:07.072607 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7\": container with ID starting with d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7 not found: ID does not exist" containerID="d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.072818 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7"} err="failed to get container status \"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7\": rpc error: code = NotFound desc = could not find container \"d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7\": container with ID starting with d6878affc892cb5b4271c45f8cee8ff42b9d8fdf01299c415ba92d89ce5a1fb7 not found: ID does not exist" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.072960 4872 scope.go:117] "RemoveContainer" containerID="66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862" Oct 09 08:56:07 crc kubenswrapper[4872]: E1009 08:56:07.073416 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862\": container with ID starting with 66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862 not found: ID does not exist" containerID="66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.073450 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862"} err="failed to get container status \"66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862\": rpc error: code = NotFound desc = could not find container \"66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862\": container with ID starting with 66c9fa5f48881cfbe1e87733122c78e5f95c95c4f20d331a582463ffd4693862 not found: ID does not exist" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.073466 4872 scope.go:117] "RemoveContainer" containerID="0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff" Oct 09 08:56:07 crc kubenswrapper[4872]: E1009 08:56:07.073696 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff\": container with ID starting with 0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff not found: ID does not exist" containerID="0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff" Oct 09 08:56:07 crc kubenswrapper[4872]: I1009 08:56:07.073734 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff"} err="failed to get container status \"0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff\": rpc error: code = NotFound desc = could not find container \"0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff\": container with ID starting with 0111e00455bf571bca60db9c3f14da43243e7a49201b3ec903b071e9399617ff not found: ID does not exist" Oct 09 08:56:08 crc kubenswrapper[4872]: I1009 08:56:08.477454 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77415b07-3938-4072-805b-64fff6789e72" path="/var/lib/kubelet/pods/77415b07-3938-4072-805b-64fff6789e72/volumes" Oct 09 08:56:17 crc kubenswrapper[4872]: I1009 08:56:17.462231 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:56:17 crc kubenswrapper[4872]: E1009 08:56:17.463191 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:56:29 crc kubenswrapper[4872]: I1009 08:56:29.461709 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:56:29 crc kubenswrapper[4872]: E1009 08:56:29.463060 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:56:43 crc kubenswrapper[4872]: I1009 08:56:43.462248 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:56:43 crc kubenswrapper[4872]: E1009 08:56:43.463211 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:56:58 crc kubenswrapper[4872]: I1009 08:56:58.461609 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:56:58 crc kubenswrapper[4872]: E1009 08:56:58.462465 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:57:11 crc kubenswrapper[4872]: I1009 08:57:11.613520 4872 generic.go:334] "Generic (PLEG): container finished" podID="80ca0a43-215b-42e4-8278-c7ded62e8080" containerID="e481328e7ea274287cc6995242b09c994010b44f9c5700380a80375bf3987976" exitCode=0 Oct 09 08:57:11 crc kubenswrapper[4872]: I1009 08:57:11.613617 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" event={"ID":"80ca0a43-215b-42e4-8278-c7ded62e8080","Type":"ContainerDied","Data":"e481328e7ea274287cc6995242b09c994010b44f9c5700380a80375bf3987976"} Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.074970 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.232493 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory\") pod \"80ca0a43-215b-42e4-8278-c7ded62e8080\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.232602 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0\") pod \"80ca0a43-215b-42e4-8278-c7ded62e8080\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.232688 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle\") pod \"80ca0a43-215b-42e4-8278-c7ded62e8080\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.232877 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key\") pod \"80ca0a43-215b-42e4-8278-c7ded62e8080\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.232973 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfkjt\" (UniqueName: \"kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt\") pod \"80ca0a43-215b-42e4-8278-c7ded62e8080\" (UID: \"80ca0a43-215b-42e4-8278-c7ded62e8080\") " Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.238160 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt" (OuterVolumeSpecName: "kube-api-access-lfkjt") pod "80ca0a43-215b-42e4-8278-c7ded62e8080" (UID: "80ca0a43-215b-42e4-8278-c7ded62e8080"). InnerVolumeSpecName "kube-api-access-lfkjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.238247 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "80ca0a43-215b-42e4-8278-c7ded62e8080" (UID: "80ca0a43-215b-42e4-8278-c7ded62e8080"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.261262 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "80ca0a43-215b-42e4-8278-c7ded62e8080" (UID: "80ca0a43-215b-42e4-8278-c7ded62e8080"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.261414 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "80ca0a43-215b-42e4-8278-c7ded62e8080" (UID: "80ca0a43-215b-42e4-8278-c7ded62e8080"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.262956 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory" (OuterVolumeSpecName: "inventory") pod "80ca0a43-215b-42e4-8278-c7ded62e8080" (UID: "80ca0a43-215b-42e4-8278-c7ded62e8080"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.335999 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.336040 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfkjt\" (UniqueName: \"kubernetes.io/projected/80ca0a43-215b-42e4-8278-c7ded62e8080-kube-api-access-lfkjt\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.336056 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.336069 4872 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.336081 4872 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80ca0a43-215b-42e4-8278-c7ded62e8080-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.462274 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:57:13 crc kubenswrapper[4872]: E1009 08:57:13.462488 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.635027 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" event={"ID":"80ca0a43-215b-42e4-8278-c7ded62e8080","Type":"ContainerDied","Data":"7b7ad4c5e651ec1e157079a87ca009bc1e9f0c0c712508af59307e4eb60ad3cf"} Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.635084 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b7ad4c5e651ec1e157079a87ca009bc1e9f0c0c712508af59307e4eb60ad3cf" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.635153 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-dvntq" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.737742 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7"] Oct 09 08:57:13 crc kubenswrapper[4872]: E1009 08:57:13.738132 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="extract-utilities" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738151 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="extract-utilities" Oct 09 08:57:13 crc kubenswrapper[4872]: E1009 08:57:13.738169 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80ca0a43-215b-42e4-8278-c7ded62e8080" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738177 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="80ca0a43-215b-42e4-8278-c7ded62e8080" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 08:57:13 crc kubenswrapper[4872]: E1009 08:57:13.738189 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="extract-content" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738194 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="extract-content" Oct 09 08:57:13 crc kubenswrapper[4872]: E1009 08:57:13.738203 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="registry-server" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738209 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="registry-server" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738361 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="80ca0a43-215b-42e4-8278-c7ded62e8080" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.738384 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="77415b07-3938-4072-805b-64fff6789e72" containerName="registry-server" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.739125 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.742938 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.743007 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.743080 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.743191 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.743194 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.743243 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.745389 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.750694 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7"] Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.845859 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.845907 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.845967 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.845989 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28mt7\" (UniqueName: \"kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.846093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.846112 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.846142 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.846157 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.846194 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.947729 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948073 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948098 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28mt7\" (UniqueName: \"kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948143 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948164 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948190 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948214 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948258 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948309 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.948674 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.953416 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.953585 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.954819 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.955504 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.956352 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.956513 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.957663 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:13 crc kubenswrapper[4872]: I1009 08:57:13.968589 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28mt7\" (UniqueName: \"kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7\") pod \"nova-edpm-deployment-openstack-edpm-ipam-h2wx7\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:14 crc kubenswrapper[4872]: I1009 08:57:14.064664 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 08:57:14 crc kubenswrapper[4872]: I1009 08:57:14.641585 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7"] Oct 09 08:57:15 crc kubenswrapper[4872]: I1009 08:57:15.657621 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" event={"ID":"fc88437f-6937-47eb-bebd-514527660dd2","Type":"ContainerStarted","Data":"0f1a4bf8a64b8ecd4ed0039a7381a01277090881d40fbf5fb9018be70e78ba2f"} Oct 09 08:57:15 crc kubenswrapper[4872]: I1009 08:57:15.657983 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" event={"ID":"fc88437f-6937-47eb-bebd-514527660dd2","Type":"ContainerStarted","Data":"2481726eafe7170002dbabd19ff3d2304c53c23573f1ae9ba36c39d6a17719dd"} Oct 09 08:57:15 crc kubenswrapper[4872]: I1009 08:57:15.677752 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" podStartSLOduration=2.117568231 podStartE2EDuration="2.677736919s" podCreationTimestamp="2025-10-09 08:57:13 +0000 UTC" firstStartedPulling="2025-10-09 08:57:14.641096231 +0000 UTC m=+2392.831624857" lastFinishedPulling="2025-10-09 08:57:15.201264919 +0000 UTC m=+2393.391793545" observedRunningTime="2025-10-09 08:57:15.674921786 +0000 UTC m=+2393.865450412" watchObservedRunningTime="2025-10-09 08:57:15.677736919 +0000 UTC m=+2393.868265535" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.103037 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.105911 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.112711 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.278800 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.279160 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6zws\" (UniqueName: \"kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.279233 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.380476 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6zws\" (UniqueName: \"kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.380574 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.380675 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.381159 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.381232 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.402140 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6zws\" (UniqueName: \"kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws\") pod \"redhat-operators-g5rxl\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.426526 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:25 crc kubenswrapper[4872]: I1009 08:57:25.883103 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:26 crc kubenswrapper[4872]: I1009 08:57:26.751849 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerID="54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27" exitCode=0 Oct 09 08:57:26 crc kubenswrapper[4872]: I1009 08:57:26.751940 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerDied","Data":"54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27"} Oct 09 08:57:26 crc kubenswrapper[4872]: I1009 08:57:26.752423 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerStarted","Data":"2f20b557a1d4bc7b69232fea19ea83c27e67bc1ffe7e821f29e88cf3b789c3b9"} Oct 09 08:57:28 crc kubenswrapper[4872]: I1009 08:57:28.461824 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:57:28 crc kubenswrapper[4872]: E1009 08:57:28.462897 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:57:28 crc kubenswrapper[4872]: I1009 08:57:28.771501 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerID="2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde" exitCode=0 Oct 09 08:57:28 crc kubenswrapper[4872]: I1009 08:57:28.771682 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerDied","Data":"2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde"} Oct 09 08:57:29 crc kubenswrapper[4872]: I1009 08:57:29.784045 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerStarted","Data":"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f"} Oct 09 08:57:29 crc kubenswrapper[4872]: I1009 08:57:29.807370 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5rxl" podStartSLOduration=2.367775696 podStartE2EDuration="4.807348413s" podCreationTimestamp="2025-10-09 08:57:25 +0000 UTC" firstStartedPulling="2025-10-09 08:57:26.754195047 +0000 UTC m=+2404.944723703" lastFinishedPulling="2025-10-09 08:57:29.193767794 +0000 UTC m=+2407.384296420" observedRunningTime="2025-10-09 08:57:29.803910803 +0000 UTC m=+2407.994439469" watchObservedRunningTime="2025-10-09 08:57:29.807348413 +0000 UTC m=+2407.997877049" Oct 09 08:57:35 crc kubenswrapper[4872]: I1009 08:57:35.427422 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:35 crc kubenswrapper[4872]: I1009 08:57:35.429179 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:35 crc kubenswrapper[4872]: I1009 08:57:35.482617 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:35 crc kubenswrapper[4872]: I1009 08:57:35.916095 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:35 crc kubenswrapper[4872]: I1009 08:57:35.981408 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:37 crc kubenswrapper[4872]: I1009 08:57:37.861456 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g5rxl" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="registry-server" containerID="cri-o://632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f" gracePeriod=2 Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.322475 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.446399 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content\") pod \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.446510 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities\") pod \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.446831 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6zws\" (UniqueName: \"kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws\") pod \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\" (UID: \"76a3c0ff-5abf-402d-bab8-f6088d9ca357\") " Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.448322 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities" (OuterVolumeSpecName: "utilities") pod "76a3c0ff-5abf-402d-bab8-f6088d9ca357" (UID: "76a3c0ff-5abf-402d-bab8-f6088d9ca357"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.453823 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws" (OuterVolumeSpecName: "kube-api-access-c6zws") pod "76a3c0ff-5abf-402d-bab8-f6088d9ca357" (UID: "76a3c0ff-5abf-402d-bab8-f6088d9ca357"). InnerVolumeSpecName "kube-api-access-c6zws". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.549091 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.549511 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6zws\" (UniqueName: \"kubernetes.io/projected/76a3c0ff-5abf-402d-bab8-f6088d9ca357-kube-api-access-c6zws\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.888523 4872 generic.go:334] "Generic (PLEG): container finished" podID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerID="632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f" exitCode=0 Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.888615 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerDied","Data":"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f"} Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.888703 4872 scope.go:117] "RemoveContainer" containerID="632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.888738 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5rxl" event={"ID":"76a3c0ff-5abf-402d-bab8-f6088d9ca357","Type":"ContainerDied","Data":"2f20b557a1d4bc7b69232fea19ea83c27e67bc1ffe7e821f29e88cf3b789c3b9"} Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.888914 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5rxl" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.942020 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76a3c0ff-5abf-402d-bab8-f6088d9ca357" (UID: "76a3c0ff-5abf-402d-bab8-f6088d9ca357"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.942667 4872 scope.go:117] "RemoveContainer" containerID="2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.962269 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76a3c0ff-5abf-402d-bab8-f6088d9ca357-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 08:57:38 crc kubenswrapper[4872]: I1009 08:57:38.964987 4872 scope.go:117] "RemoveContainer" containerID="54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.019918 4872 scope.go:117] "RemoveContainer" containerID="632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f" Oct 09 08:57:39 crc kubenswrapper[4872]: E1009 08:57:39.020441 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f\": container with ID starting with 632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f not found: ID does not exist" containerID="632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.020473 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f"} err="failed to get container status \"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f\": rpc error: code = NotFound desc = could not find container \"632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f\": container with ID starting with 632fa8f39053383218147c0ca39d4235b5851ef133a40827d54e699250e0e07f not found: ID does not exist" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.020496 4872 scope.go:117] "RemoveContainer" containerID="2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde" Oct 09 08:57:39 crc kubenswrapper[4872]: E1009 08:57:39.020986 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde\": container with ID starting with 2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde not found: ID does not exist" containerID="2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.021027 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde"} err="failed to get container status \"2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde\": rpc error: code = NotFound desc = could not find container \"2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde\": container with ID starting with 2821c47d2da12e627441d06ba12c37f6a7afbd662ba16e3984a57370293ffcde not found: ID does not exist" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.021048 4872 scope.go:117] "RemoveContainer" containerID="54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27" Oct 09 08:57:39 crc kubenswrapper[4872]: E1009 08:57:39.021362 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27\": container with ID starting with 54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27 not found: ID does not exist" containerID="54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.021384 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27"} err="failed to get container status \"54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27\": rpc error: code = NotFound desc = could not find container \"54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27\": container with ID starting with 54a6370bd767ebad0890f920be79a7bf2848faedb39164ca99c596cc60b12f27 not found: ID does not exist" Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.232382 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:39 crc kubenswrapper[4872]: I1009 08:57:39.240446 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g5rxl"] Oct 09 08:57:40 crc kubenswrapper[4872]: I1009 08:57:40.476125 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" path="/var/lib/kubelet/pods/76a3c0ff-5abf-402d-bab8-f6088d9ca357/volumes" Oct 09 08:57:43 crc kubenswrapper[4872]: I1009 08:57:43.462195 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:57:43 crc kubenswrapper[4872]: E1009 08:57:43.463054 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:57:54 crc kubenswrapper[4872]: I1009 08:57:54.462012 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:57:54 crc kubenswrapper[4872]: E1009 08:57:54.463365 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:58:06 crc kubenswrapper[4872]: I1009 08:58:06.461336 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:58:06 crc kubenswrapper[4872]: E1009 08:58:06.462341 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:58:20 crc kubenswrapper[4872]: I1009 08:58:20.462902 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:58:20 crc kubenswrapper[4872]: E1009 08:58:20.466915 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:58:31 crc kubenswrapper[4872]: I1009 08:58:31.462681 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:58:31 crc kubenswrapper[4872]: E1009 08:58:31.463670 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:58:45 crc kubenswrapper[4872]: I1009 08:58:45.462615 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:58:45 crc kubenswrapper[4872]: E1009 08:58:45.464008 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:58:57 crc kubenswrapper[4872]: I1009 08:58:57.462483 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:58:57 crc kubenswrapper[4872]: E1009 08:58:57.463460 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:59:08 crc kubenswrapper[4872]: I1009 08:59:08.461725 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:59:08 crc kubenswrapper[4872]: E1009 08:59:08.462563 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:59:20 crc kubenswrapper[4872]: I1009 08:59:20.462068 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:59:20 crc kubenswrapper[4872]: E1009 08:59:20.462859 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:59:32 crc kubenswrapper[4872]: I1009 08:59:32.467804 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:59:32 crc kubenswrapper[4872]: E1009 08:59:32.468390 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:59:44 crc kubenswrapper[4872]: I1009 08:59:44.462828 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:59:44 crc kubenswrapper[4872]: E1009 08:59:44.466299 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 08:59:55 crc kubenswrapper[4872]: I1009 08:59:55.461442 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 08:59:55 crc kubenswrapper[4872]: E1009 08:59:55.462111 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.159916 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh"] Oct 09 09:00:00 crc kubenswrapper[4872]: E1009 09:00:00.161857 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="extract-utilities" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.161887 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="extract-utilities" Oct 09 09:00:00 crc kubenswrapper[4872]: E1009 09:00:00.161906 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="extract-content" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.161916 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="extract-content" Oct 09 09:00:00 crc kubenswrapper[4872]: E1009 09:00:00.161945 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="registry-server" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.161953 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="registry-server" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.162190 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="76a3c0ff-5abf-402d-bab8-f6088d9ca357" containerName="registry-server" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.163023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.166418 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.167003 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.177352 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh"] Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.296699 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.296807 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v94l7\" (UniqueName: \"kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.296870 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.398671 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.398780 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v94l7\" (UniqueName: \"kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.398818 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.400533 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.409101 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.419013 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v94l7\" (UniqueName: \"kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7\") pod \"collect-profiles-29333340-nt4hh\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.489097 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:00 crc kubenswrapper[4872]: I1009 09:00:00.930656 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh"] Oct 09 09:00:01 crc kubenswrapper[4872]: I1009 09:00:01.270934 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" event={"ID":"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83","Type":"ContainerStarted","Data":"3e5056e7361655efdf2426a306f94d9af43d2531dc1ba928801ab485e94c4598"} Oct 09 09:00:01 crc kubenswrapper[4872]: I1009 09:00:01.271014 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" event={"ID":"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83","Type":"ContainerStarted","Data":"64c7c296fd10952d6f727379f463f14a2b81dc51e87e0c6905fc74c90367c315"} Oct 09 09:00:01 crc kubenswrapper[4872]: I1009 09:00:01.287362 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" podStartSLOduration=1.287334484 podStartE2EDuration="1.287334484s" podCreationTimestamp="2025-10-09 09:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 09:00:01.284277854 +0000 UTC m=+2559.474806480" watchObservedRunningTime="2025-10-09 09:00:01.287334484 +0000 UTC m=+2559.477863110" Oct 09 09:00:02 crc kubenswrapper[4872]: I1009 09:00:02.290570 4872 generic.go:334] "Generic (PLEG): container finished" podID="a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" containerID="3e5056e7361655efdf2426a306f94d9af43d2531dc1ba928801ab485e94c4598" exitCode=0 Oct 09 09:00:02 crc kubenswrapper[4872]: I1009 09:00:02.291068 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" event={"ID":"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83","Type":"ContainerDied","Data":"3e5056e7361655efdf2426a306f94d9af43d2531dc1ba928801ab485e94c4598"} Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.660471 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.758177 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v94l7\" (UniqueName: \"kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7\") pod \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.758241 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume\") pod \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.758274 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume\") pod \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\" (UID: \"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83\") " Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.759160 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume" (OuterVolumeSpecName: "config-volume") pod "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" (UID: "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.784034 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" (UID: "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.784674 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7" (OuterVolumeSpecName: "kube-api-access-v94l7") pod "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" (UID: "a8c3269e-f8c2-4486-9da5-0f6efdfd1d83"). InnerVolumeSpecName "kube-api-access-v94l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.860395 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v94l7\" (UniqueName: \"kubernetes.io/projected/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-kube-api-access-v94l7\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.860426 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:03 crc kubenswrapper[4872]: I1009 09:00:03.860435 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a8c3269e-f8c2-4486-9da5-0f6efdfd1d83-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.313634 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" event={"ID":"a8c3269e-f8c2-4486-9da5-0f6efdfd1d83","Type":"ContainerDied","Data":"64c7c296fd10952d6f727379f463f14a2b81dc51e87e0c6905fc74c90367c315"} Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.313882 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64c7c296fd10952d6f727379f463f14a2b81dc51e87e0c6905fc74c90367c315" Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.313709 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333340-nt4hh" Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.361027 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl"] Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.369508 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333295-tptvl"] Oct 09 09:00:04 crc kubenswrapper[4872]: I1009 09:00:04.474509 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b03f0070-47ed-455e-b4b8-7a86fa5818af" path="/var/lib/kubelet/pods/b03f0070-47ed-455e-b4b8-7a86fa5818af/volumes" Oct 09 09:00:10 crc kubenswrapper[4872]: I1009 09:00:10.461981 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 09:00:11 crc kubenswrapper[4872]: I1009 09:00:11.392428 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f"} Oct 09 09:00:15 crc kubenswrapper[4872]: I1009 09:00:15.436216 4872 generic.go:334] "Generic (PLEG): container finished" podID="fc88437f-6937-47eb-bebd-514527660dd2" containerID="0f1a4bf8a64b8ecd4ed0039a7381a01277090881d40fbf5fb9018be70e78ba2f" exitCode=0 Oct 09 09:00:15 crc kubenswrapper[4872]: I1009 09:00:15.436295 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" event={"ID":"fc88437f-6937-47eb-bebd-514527660dd2","Type":"ContainerDied","Data":"0f1a4bf8a64b8ecd4ed0039a7381a01277090881d40fbf5fb9018be70e78ba2f"} Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.836038 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930050 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930249 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930358 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28mt7\" (UniqueName: \"kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930387 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930427 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930494 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.930519 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.931411 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.931513 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key\") pod \"fc88437f-6937-47eb-bebd-514527660dd2\" (UID: \"fc88437f-6937-47eb-bebd-514527660dd2\") " Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.939854 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7" (OuterVolumeSpecName: "kube-api-access-28mt7") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "kube-api-access-28mt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.940769 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.961576 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.961605 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.961579 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.962667 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.963506 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.967143 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:16 crc kubenswrapper[4872]: I1009 09:00:16.968454 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory" (OuterVolumeSpecName: "inventory") pod "fc88437f-6937-47eb-bebd-514527660dd2" (UID: "fc88437f-6937-47eb-bebd-514527660dd2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035041 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28mt7\" (UniqueName: \"kubernetes.io/projected/fc88437f-6937-47eb-bebd-514527660dd2-kube-api-access-28mt7\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035077 4872 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035090 4872 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035174 4872 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035233 4872 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035247 4872 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035260 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035272 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/fc88437f-6937-47eb-bebd-514527660dd2-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.035286 4872 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/fc88437f-6937-47eb-bebd-514527660dd2-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.451682 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" event={"ID":"fc88437f-6937-47eb-bebd-514527660dd2","Type":"ContainerDied","Data":"2481726eafe7170002dbabd19ff3d2304c53c23573f1ae9ba36c39d6a17719dd"} Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.451737 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-h2wx7" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.451740 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2481726eafe7170002dbabd19ff3d2304c53c23573f1ae9ba36c39d6a17719dd" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.559107 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8"] Oct 09 09:00:17 crc kubenswrapper[4872]: E1009 09:00:17.559579 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc88437f-6937-47eb-bebd-514527660dd2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.559602 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc88437f-6937-47eb-bebd-514527660dd2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 09 09:00:17 crc kubenswrapper[4872]: E1009 09:00:17.559631 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" containerName="collect-profiles" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.559731 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" containerName="collect-profiles" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.559986 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a8c3269e-f8c2-4486-9da5-0f6efdfd1d83" containerName="collect-profiles" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.560025 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc88437f-6937-47eb-bebd-514527660dd2" containerName="nova-edpm-deployment-openstack-edpm-ipam" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.560803 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.563758 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.564541 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-5n5kd" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.564599 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.564996 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.565544 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.567729 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8"] Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.647800 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.647975 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.648012 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.648081 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.648100 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.648121 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.648250 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rn59l\" (UniqueName: \"kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750457 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750510 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750543 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750597 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rn59l\" (UniqueName: \"kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750629 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750773 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.750930 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.756860 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.756934 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.757400 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.757484 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.758337 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.758744 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.772166 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rn59l\" (UniqueName: \"kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-258q8\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:17 crc kubenswrapper[4872]: I1009 09:00:17.877862 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:00:18 crc kubenswrapper[4872]: I1009 09:00:18.412800 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8"] Oct 09 09:00:18 crc kubenswrapper[4872]: I1009 09:00:18.474574 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" event={"ID":"55becb4c-820e-4d84-b7c3-e3b2bddc84a0","Type":"ContainerStarted","Data":"8a304a192976b884ecddf21bab1343562b38e590138db74c43d4b291b5313e10"} Oct 09 09:00:19 crc kubenswrapper[4872]: I1009 09:00:19.471926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" event={"ID":"55becb4c-820e-4d84-b7c3-e3b2bddc84a0","Type":"ContainerStarted","Data":"55e1d4071da8906f05c8d872b6b00fab9799f9ae9698a9684fc8e8a59a5aeba0"} Oct 09 09:00:19 crc kubenswrapper[4872]: I1009 09:00:19.499108 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" podStartSLOduration=1.880971564 podStartE2EDuration="2.499090945s" podCreationTimestamp="2025-10-09 09:00:17 +0000 UTC" firstStartedPulling="2025-10-09 09:00:18.417071367 +0000 UTC m=+2576.607599993" lastFinishedPulling="2025-10-09 09:00:19.035190748 +0000 UTC m=+2577.225719374" observedRunningTime="2025-10-09 09:00:19.486443164 +0000 UTC m=+2577.676971790" watchObservedRunningTime="2025-10-09 09:00:19.499090945 +0000 UTC m=+2577.689619571" Oct 09 09:00:48 crc kubenswrapper[4872]: I1009 09:00:48.070018 4872 scope.go:117] "RemoveContainer" containerID="eb5dd41a113c1b3357182b82302eec5661dc52959845534b7b565ae194b27be8" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.146628 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29333341-v6cn6"] Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.149091 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.157970 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29333341-v6cn6"] Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.211089 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.211164 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.211195 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.211460 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pnnt\" (UniqueName: \"kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.313462 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.313581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.313740 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pnnt\" (UniqueName: \"kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.313786 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.321337 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.322362 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.331210 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.332035 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pnnt\" (UniqueName: \"kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt\") pod \"keystone-cron-29333341-v6cn6\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.484131 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:00 crc kubenswrapper[4872]: I1009 09:01:00.939747 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29333341-v6cn6"] Oct 09 09:01:01 crc kubenswrapper[4872]: I1009 09:01:01.879096 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333341-v6cn6" event={"ID":"3c82d19d-ec17-47d9-aace-a14cfb593965","Type":"ContainerStarted","Data":"18e3117ef792f8a7a3aabf584bdd812b2f83eab6d706cb6aaae8d85855390763"} Oct 09 09:01:01 crc kubenswrapper[4872]: I1009 09:01:01.879436 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333341-v6cn6" event={"ID":"3c82d19d-ec17-47d9-aace-a14cfb593965","Type":"ContainerStarted","Data":"49c5c5b80895625038aacd6ca678a56d5ea623e15b2f46623aba671e4dc1c7d2"} Oct 09 09:01:01 crc kubenswrapper[4872]: I1009 09:01:01.908726 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29333341-v6cn6" podStartSLOduration=1.90870198 podStartE2EDuration="1.90870198s" podCreationTimestamp="2025-10-09 09:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 09:01:01.896852012 +0000 UTC m=+2620.087380738" watchObservedRunningTime="2025-10-09 09:01:01.90870198 +0000 UTC m=+2620.099230636" Oct 09 09:01:03 crc kubenswrapper[4872]: I1009 09:01:03.898590 4872 generic.go:334] "Generic (PLEG): container finished" podID="3c82d19d-ec17-47d9-aace-a14cfb593965" containerID="18e3117ef792f8a7a3aabf584bdd812b2f83eab6d706cb6aaae8d85855390763" exitCode=0 Oct 09 09:01:03 crc kubenswrapper[4872]: I1009 09:01:03.898681 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333341-v6cn6" event={"ID":"3c82d19d-ec17-47d9-aace-a14cfb593965","Type":"ContainerDied","Data":"18e3117ef792f8a7a3aabf584bdd812b2f83eab6d706cb6aaae8d85855390763"} Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.210623 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.317227 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data\") pod \"3c82d19d-ec17-47d9-aace-a14cfb593965\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.317348 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle\") pod \"3c82d19d-ec17-47d9-aace-a14cfb593965\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.317385 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys\") pod \"3c82d19d-ec17-47d9-aace-a14cfb593965\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.317530 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pnnt\" (UniqueName: \"kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt\") pod \"3c82d19d-ec17-47d9-aace-a14cfb593965\" (UID: \"3c82d19d-ec17-47d9-aace-a14cfb593965\") " Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.322886 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "3c82d19d-ec17-47d9-aace-a14cfb593965" (UID: "3c82d19d-ec17-47d9-aace-a14cfb593965"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.325770 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt" (OuterVolumeSpecName: "kube-api-access-4pnnt") pod "3c82d19d-ec17-47d9-aace-a14cfb593965" (UID: "3c82d19d-ec17-47d9-aace-a14cfb593965"). InnerVolumeSpecName "kube-api-access-4pnnt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.351075 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3c82d19d-ec17-47d9-aace-a14cfb593965" (UID: "3c82d19d-ec17-47d9-aace-a14cfb593965"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.395408 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data" (OuterVolumeSpecName: "config-data") pod "3c82d19d-ec17-47d9-aace-a14cfb593965" (UID: "3c82d19d-ec17-47d9-aace-a14cfb593965"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.420286 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pnnt\" (UniqueName: \"kubernetes.io/projected/3c82d19d-ec17-47d9-aace-a14cfb593965-kube-api-access-4pnnt\") on node \"crc\" DevicePath \"\"" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.420315 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.420324 4872 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.420332 4872 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/3c82d19d-ec17-47d9-aace-a14cfb593965-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.917507 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29333341-v6cn6" event={"ID":"3c82d19d-ec17-47d9-aace-a14cfb593965","Type":"ContainerDied","Data":"49c5c5b80895625038aacd6ca678a56d5ea623e15b2f46623aba671e4dc1c7d2"} Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.917562 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49c5c5b80895625038aacd6ca678a56d5ea623e15b2f46623aba671e4dc1c7d2" Oct 09 09:01:05 crc kubenswrapper[4872]: I1009 09:01:05.917570 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29333341-v6cn6" Oct 09 09:01:05 crc kubenswrapper[4872]: E1009 09:01:05.955782 4872 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3c82d19d_ec17_47d9_aace_a14cfb593965.slice\": RecentStats: unable to find data in memory cache]" Oct 09 09:02:35 crc kubenswrapper[4872]: I1009 09:02:35.827451 4872 generic.go:334] "Generic (PLEG): container finished" podID="55becb4c-820e-4d84-b7c3-e3b2bddc84a0" containerID="55e1d4071da8906f05c8d872b6b00fab9799f9ae9698a9684fc8e8a59a5aeba0" exitCode=0 Oct 09 09:02:35 crc kubenswrapper[4872]: I1009 09:02:35.827545 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" event={"ID":"55becb4c-820e-4d84-b7c3-e3b2bddc84a0","Type":"ContainerDied","Data":"55e1d4071da8906f05c8d872b6b00fab9799f9ae9698a9684fc8e8a59a5aeba0"} Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.372256 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442548 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442708 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442749 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rn59l\" (UniqueName: \"kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442827 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442869 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442887 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.442956 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle\") pod \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\" (UID: \"55becb4c-820e-4d84-b7c3-e3b2bddc84a0\") " Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.449763 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l" (OuterVolumeSpecName: "kube-api-access-rn59l") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "kube-api-access-rn59l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.450399 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.480418 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.482263 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.483355 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.490298 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory" (OuterVolumeSpecName: "inventory") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.493674 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "55becb4c-820e-4d84-b7c3-e3b2bddc84a0" (UID: "55becb4c-820e-4d84-b7c3-e3b2bddc84a0"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.544889 4872 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.544944 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.544957 4872 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.544973 4872 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.544987 4872 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.545003 4872 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-inventory\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.545014 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rn59l\" (UniqueName: \"kubernetes.io/projected/55becb4c-820e-4d84-b7c3-e3b2bddc84a0-kube-api-access-rn59l\") on node \"crc\" DevicePath \"\"" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.848773 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" event={"ID":"55becb4c-820e-4d84-b7c3-e3b2bddc84a0","Type":"ContainerDied","Data":"8a304a192976b884ecddf21bab1343562b38e590138db74c43d4b291b5313e10"} Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.848824 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a304a192976b884ecddf21bab1343562b38e590138db74c43d4b291b5313e10" Oct 09 09:02:37 crc kubenswrapper[4872]: I1009 09:02:37.848830 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-258q8" Oct 09 09:02:38 crc kubenswrapper[4872]: I1009 09:02:38.156583 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:02:38 crc kubenswrapper[4872]: I1009 09:02:38.157007 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:03:08 crc kubenswrapper[4872]: I1009 09:03:08.156413 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:03:08 crc kubenswrapper[4872]: I1009 09:03:08.157114 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.082168 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Oct 09 09:03:21 crc kubenswrapper[4872]: E1009 09:03:21.083158 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c82d19d-ec17-47d9-aace-a14cfb593965" containerName="keystone-cron" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.083175 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c82d19d-ec17-47d9-aace-a14cfb593965" containerName="keystone-cron" Oct 09 09:03:21 crc kubenswrapper[4872]: E1009 09:03:21.083196 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55becb4c-820e-4d84-b7c3-e3b2bddc84a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.083206 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="55becb4c-820e-4d84-b7c3-e3b2bddc84a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.083453 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="55becb4c-820e-4d84-b7c3-e3b2bddc84a0" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.083472 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c82d19d-ec17-47d9-aace-a14cfb593965" containerName="keystone-cron" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.084274 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.087419 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.087568 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.087627 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.089457 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2fckw" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.092076 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186327 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186378 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186426 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vhpq\" (UniqueName: \"kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186472 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186548 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186630 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186699 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186728 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.186807 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.289002 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.289095 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vhpq\" (UniqueName: \"kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.289180 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.289217 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.289281 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.290070 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.290149 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.290243 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.290582 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.290843 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.291261 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.291825 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.291914 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.292360 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.297062 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.304463 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.304617 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.309898 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vhpq\" (UniqueName: \"kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.324006 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"tempest-tests-tempest\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.404379 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.903975 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Oct 09 09:03:21 crc kubenswrapper[4872]: I1009 09:03:21.914405 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 09:03:22 crc kubenswrapper[4872]: I1009 09:03:22.258513 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efed60a8-77a0-4635-8ea2-46f3400b9090","Type":"ContainerStarted","Data":"23b1367df2bde98b4de21401a9f86e89aa3052400bfc5b3e59af96463a9ef8e5"} Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.156818 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.157329 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.157366 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.157960 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.158010 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f" gracePeriod=600 Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.440094 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f" exitCode=0 Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.440133 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f"} Oct 09 09:03:38 crc kubenswrapper[4872]: I1009 09:03:38.440164 4872 scope.go:117] "RemoveContainer" containerID="dfbb8f0b770daec5381534fde415f7c12f745553531b9d62a4b27fd5bac5ef8a" Oct 09 09:03:54 crc kubenswrapper[4872]: E1009 09:03:54.523687 4872 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Oct 09 09:03:54 crc kubenswrapper[4872]: E1009 09:03:54.524487 4872 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6vhpq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(efed60a8-77a0-4635-8ea2-46f3400b9090): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 09 09:03:54 crc kubenswrapper[4872]: E1009 09:03:54.525662 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="efed60a8-77a0-4635-8ea2-46f3400b9090" Oct 09 09:03:54 crc kubenswrapper[4872]: E1009 09:03:54.593358 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="efed60a8-77a0-4635-8ea2-46f3400b9090" Oct 09 09:03:55 crc kubenswrapper[4872]: I1009 09:03:55.601884 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2"} Oct 09 09:04:11 crc kubenswrapper[4872]: I1009 09:04:11.096511 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Oct 09 09:04:12 crc kubenswrapper[4872]: I1009 09:04:12.765117 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efed60a8-77a0-4635-8ea2-46f3400b9090","Type":"ContainerStarted","Data":"2510a36c00800f7f7dc1123d164256f3f304093b69293f95ab77e1f85d648116"} Oct 09 09:04:12 crc kubenswrapper[4872]: I1009 09:04:12.793515 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=3.613386087 podStartE2EDuration="52.793496292s" podCreationTimestamp="2025-10-09 09:03:20 +0000 UTC" firstStartedPulling="2025-10-09 09:03:21.914172406 +0000 UTC m=+2760.104701032" lastFinishedPulling="2025-10-09 09:04:11.094282611 +0000 UTC m=+2809.284811237" observedRunningTime="2025-10-09 09:04:12.788839815 +0000 UTC m=+2810.979368471" watchObservedRunningTime="2025-10-09 09:04:12.793496292 +0000 UTC m=+2810.984024918" Oct 09 09:06:08 crc kubenswrapper[4872]: I1009 09:06:08.156940 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:06:08 crc kubenswrapper[4872]: I1009 09:06:08.158803 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:06:38 crc kubenswrapper[4872]: I1009 09:06:38.157699 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:06:38 crc kubenswrapper[4872]: I1009 09:06:38.158161 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.638469 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.642350 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.657036 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.798924 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9jtr\" (UniqueName: \"kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.798979 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.799009 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.901566 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9jtr\" (UniqueName: \"kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.901674 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.901717 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.902888 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.902915 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.928855 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9jtr\" (UniqueName: \"kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr\") pod \"redhat-marketplace-zkhxq\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:40 crc kubenswrapper[4872]: I1009 09:06:40.970284 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:41 crc kubenswrapper[4872]: I1009 09:06:41.443830 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:42 crc kubenswrapper[4872]: I1009 09:06:42.184456 4872 generic.go:334] "Generic (PLEG): container finished" podID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerID="f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15" exitCode=0 Oct 09 09:06:42 crc kubenswrapper[4872]: I1009 09:06:42.184546 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerDied","Data":"f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15"} Oct 09 09:06:42 crc kubenswrapper[4872]: I1009 09:06:42.185715 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerStarted","Data":"1c53774bfdbe3ca91caed1ecba7acd5afcfa39553f203116b519efd8c45fc539"} Oct 09 09:06:44 crc kubenswrapper[4872]: I1009 09:06:44.205035 4872 generic.go:334] "Generic (PLEG): container finished" podID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerID="924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5" exitCode=0 Oct 09 09:06:44 crc kubenswrapper[4872]: I1009 09:06:44.205145 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerDied","Data":"924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5"} Oct 09 09:06:45 crc kubenswrapper[4872]: I1009 09:06:45.215788 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerStarted","Data":"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4"} Oct 09 09:06:45 crc kubenswrapper[4872]: I1009 09:06:45.238509 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zkhxq" podStartSLOduration=2.727271828 podStartE2EDuration="5.238487866s" podCreationTimestamp="2025-10-09 09:06:40 +0000 UTC" firstStartedPulling="2025-10-09 09:06:42.18729959 +0000 UTC m=+2960.377828216" lastFinishedPulling="2025-10-09 09:06:44.698515608 +0000 UTC m=+2962.889044254" observedRunningTime="2025-10-09 09:06:45.232596074 +0000 UTC m=+2963.423124700" watchObservedRunningTime="2025-10-09 09:06:45.238487866 +0000 UTC m=+2963.429016492" Oct 09 09:06:50 crc kubenswrapper[4872]: I1009 09:06:50.971242 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:50 crc kubenswrapper[4872]: I1009 09:06:50.971837 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:51 crc kubenswrapper[4872]: I1009 09:06:51.023583 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:51 crc kubenswrapper[4872]: I1009 09:06:51.331553 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:51 crc kubenswrapper[4872]: I1009 09:06:51.377388 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.288767 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zkhxq" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="registry-server" containerID="cri-o://b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4" gracePeriod=2 Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.768871 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.850824 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities\") pod \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.850905 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content\") pod \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.850995 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b9jtr\" (UniqueName: \"kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr\") pod \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\" (UID: \"fdd72a7c-304a-4474-bc3a-ed57cdca5e82\") " Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.851985 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities" (OuterVolumeSpecName: "utilities") pod "fdd72a7c-304a-4474-bc3a-ed57cdca5e82" (UID: "fdd72a7c-304a-4474-bc3a-ed57cdca5e82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.861089 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr" (OuterVolumeSpecName: "kube-api-access-b9jtr") pod "fdd72a7c-304a-4474-bc3a-ed57cdca5e82" (UID: "fdd72a7c-304a-4474-bc3a-ed57cdca5e82"). InnerVolumeSpecName "kube-api-access-b9jtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.867628 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fdd72a7c-304a-4474-bc3a-ed57cdca5e82" (UID: "fdd72a7c-304a-4474-bc3a-ed57cdca5e82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.953360 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.953405 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:06:53 crc kubenswrapper[4872]: I1009 09:06:53.953419 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b9jtr\" (UniqueName: \"kubernetes.io/projected/fdd72a7c-304a-4474-bc3a-ed57cdca5e82-kube-api-access-b9jtr\") on node \"crc\" DevicePath \"\"" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.299594 4872 generic.go:334] "Generic (PLEG): container finished" podID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerID="b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4" exitCode=0 Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.299630 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerDied","Data":"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4"} Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.299673 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zkhxq" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.299691 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zkhxq" event={"ID":"fdd72a7c-304a-4474-bc3a-ed57cdca5e82","Type":"ContainerDied","Data":"1c53774bfdbe3ca91caed1ecba7acd5afcfa39553f203116b519efd8c45fc539"} Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.299714 4872 scope.go:117] "RemoveContainer" containerID="b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.337542 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.337729 4872 scope.go:117] "RemoveContainer" containerID="924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.351121 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zkhxq"] Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.364147 4872 scope.go:117] "RemoveContainer" containerID="f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.410599 4872 scope.go:117] "RemoveContainer" containerID="b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4" Oct 09 09:06:54 crc kubenswrapper[4872]: E1009 09:06:54.411209 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4\": container with ID starting with b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4 not found: ID does not exist" containerID="b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.411260 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4"} err="failed to get container status \"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4\": rpc error: code = NotFound desc = could not find container \"b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4\": container with ID starting with b4c0106d888ec39d72de3958250b7eb4e391202a738489ba6932da3b27fdf3c4 not found: ID does not exist" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.411287 4872 scope.go:117] "RemoveContainer" containerID="924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5" Oct 09 09:06:54 crc kubenswrapper[4872]: E1009 09:06:54.411747 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5\": container with ID starting with 924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5 not found: ID does not exist" containerID="924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.411799 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5"} err="failed to get container status \"924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5\": rpc error: code = NotFound desc = could not find container \"924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5\": container with ID starting with 924cafc2ff6225feef9d7cee8cff6482542cb7d46cdc739e2ec4221dc8ff9db5 not found: ID does not exist" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.411834 4872 scope.go:117] "RemoveContainer" containerID="f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15" Oct 09 09:06:54 crc kubenswrapper[4872]: E1009 09:06:54.412160 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15\": container with ID starting with f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15 not found: ID does not exist" containerID="f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.412189 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15"} err="failed to get container status \"f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15\": rpc error: code = NotFound desc = could not find container \"f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15\": container with ID starting with f5abb8969bcb42bc451ba7b1007a2b7a30c277b2f7e71a379e2dabf5ba4e0f15 not found: ID does not exist" Oct 09 09:06:54 crc kubenswrapper[4872]: I1009 09:06:54.475801 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" path="/var/lib/kubelet/pods/fdd72a7c-304a-4474-bc3a-ed57cdca5e82/volumes" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.157470 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.158094 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.158176 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.158783 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.158876 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" gracePeriod=600 Oct 09 09:07:08 crc kubenswrapper[4872]: E1009 09:07:08.297081 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.463556 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" exitCode=0 Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.479947 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2"} Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.480007 4872 scope.go:117] "RemoveContainer" containerID="17e28f954b3cb3d8748e974e16fc7740b2426efee5659258f217de99be89828f" Oct 09 09:07:08 crc kubenswrapper[4872]: I1009 09:07:08.480829 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:07:08 crc kubenswrapper[4872]: E1009 09:07:08.481383 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:07:21 crc kubenswrapper[4872]: I1009 09:07:21.461502 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:07:21 crc kubenswrapper[4872]: E1009 09:07:21.462279 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:07:33 crc kubenswrapper[4872]: I1009 09:07:33.461558 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:07:33 crc kubenswrapper[4872]: E1009 09:07:33.462416 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:07:45 crc kubenswrapper[4872]: I1009 09:07:45.461518 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:07:45 crc kubenswrapper[4872]: E1009 09:07:45.462371 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:07:57 crc kubenswrapper[4872]: I1009 09:07:57.462283 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:07:57 crc kubenswrapper[4872]: E1009 09:07:57.463141 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:08:10 crc kubenswrapper[4872]: I1009 09:08:10.462809 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:08:10 crc kubenswrapper[4872]: E1009 09:08:10.463729 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.797007 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:19 crc kubenswrapper[4872]: E1009 09:08:19.797822 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="extract-utilities" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.797833 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="extract-utilities" Oct 09 09:08:19 crc kubenswrapper[4872]: E1009 09:08:19.797850 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="extract-content" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.797856 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="extract-content" Oct 09 09:08:19 crc kubenswrapper[4872]: E1009 09:08:19.797883 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="registry-server" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.797889 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="registry-server" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.798094 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdd72a7c-304a-4474-bc3a-ed57cdca5e82" containerName="registry-server" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.799369 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.811209 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.967124 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.967178 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:19 crc kubenswrapper[4872]: I1009 09:08:19.967694 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lsfk\" (UniqueName: \"kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.069705 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lsfk\" (UniqueName: \"kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.069831 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.069872 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.070383 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.070489 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.093462 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lsfk\" (UniqueName: \"kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk\") pod \"redhat-operators-dxn8l\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.122785 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:20 crc kubenswrapper[4872]: I1009 09:08:20.603991 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:21 crc kubenswrapper[4872]: I1009 09:08:21.154261 4872 generic.go:334] "Generic (PLEG): container finished" podID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerID="72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1" exitCode=0 Oct 09 09:08:21 crc kubenswrapper[4872]: I1009 09:08:21.154390 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerDied","Data":"72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1"} Oct 09 09:08:21 crc kubenswrapper[4872]: I1009 09:08:21.154582 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerStarted","Data":"956e953aeafa3b93405a8bf34837e73944dabb9bb1443816a07507bfe186155b"} Oct 09 09:08:22 crc kubenswrapper[4872]: I1009 09:08:22.462789 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:08:22 crc kubenswrapper[4872]: E1009 09:08:22.463814 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:08:23 crc kubenswrapper[4872]: I1009 09:08:23.175052 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerStarted","Data":"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2"} Oct 09 09:08:24 crc kubenswrapper[4872]: I1009 09:08:24.191274 4872 generic.go:334] "Generic (PLEG): container finished" podID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerID="760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2" exitCode=0 Oct 09 09:08:24 crc kubenswrapper[4872]: I1009 09:08:24.191340 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerDied","Data":"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2"} Oct 09 09:08:24 crc kubenswrapper[4872]: I1009 09:08:24.196287 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 09:08:25 crc kubenswrapper[4872]: I1009 09:08:25.201959 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerStarted","Data":"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49"} Oct 09 09:08:25 crc kubenswrapper[4872]: I1009 09:08:25.219582 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-dxn8l" podStartSLOduration=2.599794934 podStartE2EDuration="6.219556995s" podCreationTimestamp="2025-10-09 09:08:19 +0000 UTC" firstStartedPulling="2025-10-09 09:08:21.16057249 +0000 UTC m=+3059.351101116" lastFinishedPulling="2025-10-09 09:08:24.780334551 +0000 UTC m=+3062.970863177" observedRunningTime="2025-10-09 09:08:25.218031341 +0000 UTC m=+3063.408559997" watchObservedRunningTime="2025-10-09 09:08:25.219556995 +0000 UTC m=+3063.410085621" Oct 09 09:08:30 crc kubenswrapper[4872]: I1009 09:08:30.124087 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:30 crc kubenswrapper[4872]: I1009 09:08:30.125472 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:31 crc kubenswrapper[4872]: I1009 09:08:31.168881 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-dxn8l" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="registry-server" probeResult="failure" output=< Oct 09 09:08:31 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Oct 09 09:08:31 crc kubenswrapper[4872]: > Oct 09 09:08:35 crc kubenswrapper[4872]: I1009 09:08:35.462160 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:08:35 crc kubenswrapper[4872]: E1009 09:08:35.463230 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:08:40 crc kubenswrapper[4872]: I1009 09:08:40.194184 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:40 crc kubenswrapper[4872]: I1009 09:08:40.261922 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:40 crc kubenswrapper[4872]: I1009 09:08:40.446045 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:41 crc kubenswrapper[4872]: I1009 09:08:41.362130 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-dxn8l" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="registry-server" containerID="cri-o://2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49" gracePeriod=2 Oct 09 09:08:41 crc kubenswrapper[4872]: I1009 09:08:41.913898 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.011759 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities\") pod \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.011879 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content\") pod \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.012075 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lsfk\" (UniqueName: \"kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk\") pod \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\" (UID: \"58f18d31-dd21-4a97-8cb1-25e7548bdfdd\") " Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.013589 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities" (OuterVolumeSpecName: "utilities") pod "58f18d31-dd21-4a97-8cb1-25e7548bdfdd" (UID: "58f18d31-dd21-4a97-8cb1-25e7548bdfdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.018828 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk" (OuterVolumeSpecName: "kube-api-access-6lsfk") pod "58f18d31-dd21-4a97-8cb1-25e7548bdfdd" (UID: "58f18d31-dd21-4a97-8cb1-25e7548bdfdd"). InnerVolumeSpecName "kube-api-access-6lsfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.114090 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lsfk\" (UniqueName: \"kubernetes.io/projected/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-kube-api-access-6lsfk\") on node \"crc\" DevicePath \"\"" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.114372 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.116544 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58f18d31-dd21-4a97-8cb1-25e7548bdfdd" (UID: "58f18d31-dd21-4a97-8cb1-25e7548bdfdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.215685 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58f18d31-dd21-4a97-8cb1-25e7548bdfdd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.375037 4872 generic.go:334] "Generic (PLEG): container finished" podID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerID="2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49" exitCode=0 Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.375098 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerDied","Data":"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49"} Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.375139 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-dxn8l" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.375164 4872 scope.go:117] "RemoveContainer" containerID="2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.375147 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-dxn8l" event={"ID":"58f18d31-dd21-4a97-8cb1-25e7548bdfdd","Type":"ContainerDied","Data":"956e953aeafa3b93405a8bf34837e73944dabb9bb1443816a07507bfe186155b"} Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.399838 4872 scope.go:117] "RemoveContainer" containerID="760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.416173 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.423915 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-dxn8l"] Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.451049 4872 scope.go:117] "RemoveContainer" containerID="72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.473182 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" path="/var/lib/kubelet/pods/58f18d31-dd21-4a97-8cb1-25e7548bdfdd/volumes" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.481170 4872 scope.go:117] "RemoveContainer" containerID="2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49" Oct 09 09:08:42 crc kubenswrapper[4872]: E1009 09:08:42.481659 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49\": container with ID starting with 2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49 not found: ID does not exist" containerID="2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.481701 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49"} err="failed to get container status \"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49\": rpc error: code = NotFound desc = could not find container \"2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49\": container with ID starting with 2d3602893deb8d0af2e221c97b137ce0f388d77a9e4a5aac03f932003815ad49 not found: ID does not exist" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.481731 4872 scope.go:117] "RemoveContainer" containerID="760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2" Oct 09 09:08:42 crc kubenswrapper[4872]: E1009 09:08:42.482438 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2\": container with ID starting with 760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2 not found: ID does not exist" containerID="760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.482485 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2"} err="failed to get container status \"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2\": rpc error: code = NotFound desc = could not find container \"760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2\": container with ID starting with 760a6efe75e7c7db74a374b4d9be3678d17cb96634cd752dcd8bca0a7d3341d2 not found: ID does not exist" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.482516 4872 scope.go:117] "RemoveContainer" containerID="72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1" Oct 09 09:08:42 crc kubenswrapper[4872]: E1009 09:08:42.483030 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1\": container with ID starting with 72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1 not found: ID does not exist" containerID="72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1" Oct 09 09:08:42 crc kubenswrapper[4872]: I1009 09:08:42.483069 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1"} err="failed to get container status \"72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1\": rpc error: code = NotFound desc = could not find container \"72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1\": container with ID starting with 72b7d53c5e65b74b667921b16d6ae9815781b8d9cd3efcd7bf307108e30c8bf1 not found: ID does not exist" Oct 09 09:08:48 crc kubenswrapper[4872]: I1009 09:08:48.462012 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:08:48 crc kubenswrapper[4872]: E1009 09:08:48.463521 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:09:03 crc kubenswrapper[4872]: I1009 09:09:03.462623 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:09:03 crc kubenswrapper[4872]: E1009 09:09:03.464882 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:09:18 crc kubenswrapper[4872]: I1009 09:09:18.461418 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:09:18 crc kubenswrapper[4872]: E1009 09:09:18.462252 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:09:33 crc kubenswrapper[4872]: I1009 09:09:33.462213 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:09:33 crc kubenswrapper[4872]: E1009 09:09:33.463112 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:09:46 crc kubenswrapper[4872]: I1009 09:09:46.462668 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:09:46 crc kubenswrapper[4872]: E1009 09:09:46.464019 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.839752 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:09:54 crc kubenswrapper[4872]: E1009 09:09:54.840871 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="registry-server" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.840887 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="registry-server" Oct 09 09:09:54 crc kubenswrapper[4872]: E1009 09:09:54.840913 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="extract-utilities" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.840921 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="extract-utilities" Oct 09 09:09:54 crc kubenswrapper[4872]: E1009 09:09:54.840964 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="extract-content" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.840982 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="extract-content" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.841195 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="58f18d31-dd21-4a97-8cb1-25e7548bdfdd" containerName="registry-server" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.843157 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:54 crc kubenswrapper[4872]: I1009 09:09:54.855825 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.016738 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htmbc\" (UniqueName: \"kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.016931 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.017261 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.118555 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htmbc\" (UniqueName: \"kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.118923 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.119118 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.119617 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.119621 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.144195 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htmbc\" (UniqueName: \"kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc\") pod \"community-operators-wxkzw\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.182065 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:09:55 crc kubenswrapper[4872]: I1009 09:09:55.709568 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:09:56 crc kubenswrapper[4872]: I1009 09:09:56.078031 4872 generic.go:334] "Generic (PLEG): container finished" podID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerID="96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b" exitCode=0 Oct 09 09:09:56 crc kubenswrapper[4872]: I1009 09:09:56.078079 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerDied","Data":"96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b"} Oct 09 09:09:56 crc kubenswrapper[4872]: I1009 09:09:56.078104 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerStarted","Data":"c5c9f2a2f2fad1e16a2cd79eda7f359d12d780eaa3cc3a48f2a061d11071ec04"} Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.088728 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerStarted","Data":"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548"} Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.231864 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.234192 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.249509 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.364750 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.364850 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lz6q\" (UniqueName: \"kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.364925 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.467500 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.467581 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lz6q\" (UniqueName: \"kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.467654 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.468429 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.468444 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.492044 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lz6q\" (UniqueName: \"kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q\") pod \"certified-operators-d6b6m\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:57 crc kubenswrapper[4872]: I1009 09:09:57.553029 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:09:58 crc kubenswrapper[4872]: I1009 09:09:58.101284 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:09:58 crc kubenswrapper[4872]: I1009 09:09:58.101812 4872 generic.go:334] "Generic (PLEG): container finished" podID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerID="7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548" exitCode=0 Oct 09 09:09:58 crc kubenswrapper[4872]: W1009 09:09:58.105331 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod32de44a4_3f38_456a_9d4c_159d6c18f53e.slice/crio-453931a4a8231ea34a5513e05c3ac3bbc2f491d57a7835f20fca627d694b32bf WatchSource:0}: Error finding container 453931a4a8231ea34a5513e05c3ac3bbc2f491d57a7835f20fca627d694b32bf: Status 404 returned error can't find the container with id 453931a4a8231ea34a5513e05c3ac3bbc2f491d57a7835f20fca627d694b32bf Oct 09 09:09:58 crc kubenswrapper[4872]: I1009 09:09:58.107001 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerDied","Data":"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548"} Oct 09 09:09:59 crc kubenswrapper[4872]: I1009 09:09:59.120499 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerStarted","Data":"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f"} Oct 09 09:09:59 crc kubenswrapper[4872]: I1009 09:09:59.124312 4872 generic.go:334] "Generic (PLEG): container finished" podID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerID="e6e6b806e90ef198f82112b2333e94f45fd602cfd0175abf808f7b30751d6d10" exitCode=0 Oct 09 09:09:59 crc kubenswrapper[4872]: I1009 09:09:59.124367 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerDied","Data":"e6e6b806e90ef198f82112b2333e94f45fd602cfd0175abf808f7b30751d6d10"} Oct 09 09:09:59 crc kubenswrapper[4872]: I1009 09:09:59.124397 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerStarted","Data":"453931a4a8231ea34a5513e05c3ac3bbc2f491d57a7835f20fca627d694b32bf"} Oct 09 09:09:59 crc kubenswrapper[4872]: I1009 09:09:59.146031 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wxkzw" podStartSLOduration=2.562028749 podStartE2EDuration="5.146005376s" podCreationTimestamp="2025-10-09 09:09:54 +0000 UTC" firstStartedPulling="2025-10-09 09:09:56.085874816 +0000 UTC m=+3154.276403442" lastFinishedPulling="2025-10-09 09:09:58.669851443 +0000 UTC m=+3156.860380069" observedRunningTime="2025-10-09 09:09:59.14441667 +0000 UTC m=+3157.334945336" watchObservedRunningTime="2025-10-09 09:09:59.146005376 +0000 UTC m=+3157.336534032" Oct 09 09:10:00 crc kubenswrapper[4872]: I1009 09:10:00.137836 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerStarted","Data":"edc2cae936c92eb2b8cd4a07b3c716040a0ae3351e93f1e7cf0871d3cefca4fd"} Oct 09 09:10:00 crc kubenswrapper[4872]: I1009 09:10:00.462162 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:10:00 crc kubenswrapper[4872]: E1009 09:10:00.462468 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:10:03 crc kubenswrapper[4872]: I1009 09:10:03.169199 4872 generic.go:334] "Generic (PLEG): container finished" podID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerID="edc2cae936c92eb2b8cd4a07b3c716040a0ae3351e93f1e7cf0871d3cefca4fd" exitCode=0 Oct 09 09:10:03 crc kubenswrapper[4872]: I1009 09:10:03.169269 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerDied","Data":"edc2cae936c92eb2b8cd4a07b3c716040a0ae3351e93f1e7cf0871d3cefca4fd"} Oct 09 09:10:05 crc kubenswrapper[4872]: I1009 09:10:05.182222 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:05 crc kubenswrapper[4872]: I1009 09:10:05.182828 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:05 crc kubenswrapper[4872]: I1009 09:10:05.188327 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerStarted","Data":"d57dddf97c4c04bdcc70112f3fac34b2f07b0057128c2293241d5a2f30f47bfc"} Oct 09 09:10:05 crc kubenswrapper[4872]: I1009 09:10:05.216408 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-d6b6m" podStartSLOduration=2.437533788 podStartE2EDuration="8.216385661s" podCreationTimestamp="2025-10-09 09:09:57 +0000 UTC" firstStartedPulling="2025-10-09 09:09:59.12593597 +0000 UTC m=+3157.316464596" lastFinishedPulling="2025-10-09 09:10:04.904787843 +0000 UTC m=+3163.095316469" observedRunningTime="2025-10-09 09:10:05.210222031 +0000 UTC m=+3163.400750657" watchObservedRunningTime="2025-10-09 09:10:05.216385661 +0000 UTC m=+3163.406914287" Oct 09 09:10:05 crc kubenswrapper[4872]: I1009 09:10:05.232135 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:06 crc kubenswrapper[4872]: I1009 09:10:06.262724 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:06 crc kubenswrapper[4872]: I1009 09:10:06.627597 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:10:07 crc kubenswrapper[4872]: I1009 09:10:07.553573 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:07 crc kubenswrapper[4872]: I1009 09:10:07.553758 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:07 crc kubenswrapper[4872]: I1009 09:10:07.608196 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.217249 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wxkzw" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="registry-server" containerID="cri-o://ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f" gracePeriod=2 Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.722277 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.829156 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content\") pod \"9d884fe5-01ce-407d-acf8-10ad86c68122\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.829267 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htmbc\" (UniqueName: \"kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc\") pod \"9d884fe5-01ce-407d-acf8-10ad86c68122\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.829426 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities\") pod \"9d884fe5-01ce-407d-acf8-10ad86c68122\" (UID: \"9d884fe5-01ce-407d-acf8-10ad86c68122\") " Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.830596 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities" (OuterVolumeSpecName: "utilities") pod "9d884fe5-01ce-407d-acf8-10ad86c68122" (UID: "9d884fe5-01ce-407d-acf8-10ad86c68122"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.835113 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc" (OuterVolumeSpecName: "kube-api-access-htmbc") pod "9d884fe5-01ce-407d-acf8-10ad86c68122" (UID: "9d884fe5-01ce-407d-acf8-10ad86c68122"). InnerVolumeSpecName "kube-api-access-htmbc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.881911 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9d884fe5-01ce-407d-acf8-10ad86c68122" (UID: "9d884fe5-01ce-407d-acf8-10ad86c68122"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.931561 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.931590 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htmbc\" (UniqueName: \"kubernetes.io/projected/9d884fe5-01ce-407d-acf8-10ad86c68122-kube-api-access-htmbc\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:08 crc kubenswrapper[4872]: I1009 09:10:08.931601 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9d884fe5-01ce-407d-acf8-10ad86c68122-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.230401 4872 generic.go:334] "Generic (PLEG): container finished" podID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerID="ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f" exitCode=0 Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.230493 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerDied","Data":"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f"} Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.230531 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wxkzw" event={"ID":"9d884fe5-01ce-407d-acf8-10ad86c68122","Type":"ContainerDied","Data":"c5c9f2a2f2fad1e16a2cd79eda7f359d12d780eaa3cc3a48f2a061d11071ec04"} Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.230552 4872 scope.go:117] "RemoveContainer" containerID="ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.230495 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wxkzw" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.259959 4872 scope.go:117] "RemoveContainer" containerID="7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.294954 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.305525 4872 scope.go:117] "RemoveContainer" containerID="96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.306381 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wxkzw"] Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.346194 4872 scope.go:117] "RemoveContainer" containerID="ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f" Oct 09 09:10:09 crc kubenswrapper[4872]: E1009 09:10:09.346856 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f\": container with ID starting with ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f not found: ID does not exist" containerID="ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.346931 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f"} err="failed to get container status \"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f\": rpc error: code = NotFound desc = could not find container \"ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f\": container with ID starting with ded36bab2a3f81ca9aaf5dea7dba58805ec1e191a1cef7816d21b4aa78b7913f not found: ID does not exist" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.346976 4872 scope.go:117] "RemoveContainer" containerID="7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548" Oct 09 09:10:09 crc kubenswrapper[4872]: E1009 09:10:09.347475 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548\": container with ID starting with 7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548 not found: ID does not exist" containerID="7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.347536 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548"} err="failed to get container status \"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548\": rpc error: code = NotFound desc = could not find container \"7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548\": container with ID starting with 7fb11e355b0251decf53800cb6b1938bbcb48c9243022058c672ebf40c8d0548 not found: ID does not exist" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.347581 4872 scope.go:117] "RemoveContainer" containerID="96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b" Oct 09 09:10:09 crc kubenswrapper[4872]: E1009 09:10:09.347944 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b\": container with ID starting with 96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b not found: ID does not exist" containerID="96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b" Oct 09 09:10:09 crc kubenswrapper[4872]: I1009 09:10:09.347984 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b"} err="failed to get container status \"96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b\": rpc error: code = NotFound desc = could not find container \"96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b\": container with ID starting with 96acf7a1690032556b397dc3adcdaa05404f1d913291136909230eb1db74ef2b not found: ID does not exist" Oct 09 09:10:10 crc kubenswrapper[4872]: I1009 09:10:10.473200 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" path="/var/lib/kubelet/pods/9d884fe5-01ce-407d-acf8-10ad86c68122/volumes" Oct 09 09:10:15 crc kubenswrapper[4872]: I1009 09:10:15.462735 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:10:15 crc kubenswrapper[4872]: E1009 09:10:15.463823 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:10:17 crc kubenswrapper[4872]: I1009 09:10:17.609947 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.224503 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.225341 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-d6b6m" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="registry-server" containerID="cri-o://d57dddf97c4c04bdcc70112f3fac34b2f07b0057128c2293241d5a2f30f47bfc" gracePeriod=2 Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.358593 4872 generic.go:334] "Generic (PLEG): container finished" podID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerID="d57dddf97c4c04bdcc70112f3fac34b2f07b0057128c2293241d5a2f30f47bfc" exitCode=0 Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.358660 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerDied","Data":"d57dddf97c4c04bdcc70112f3fac34b2f07b0057128c2293241d5a2f30f47bfc"} Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.728620 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.810223 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content\") pod \"32de44a4-3f38-456a-9d4c-159d6c18f53e\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.810338 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9lz6q\" (UniqueName: \"kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q\") pod \"32de44a4-3f38-456a-9d4c-159d6c18f53e\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.810380 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities\") pod \"32de44a4-3f38-456a-9d4c-159d6c18f53e\" (UID: \"32de44a4-3f38-456a-9d4c-159d6c18f53e\") " Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.811975 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities" (OuterVolumeSpecName: "utilities") pod "32de44a4-3f38-456a-9d4c-159d6c18f53e" (UID: "32de44a4-3f38-456a-9d4c-159d6c18f53e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.816131 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q" (OuterVolumeSpecName: "kube-api-access-9lz6q") pod "32de44a4-3f38-456a-9d4c-159d6c18f53e" (UID: "32de44a4-3f38-456a-9d4c-159d6c18f53e"). InnerVolumeSpecName "kube-api-access-9lz6q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.865757 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "32de44a4-3f38-456a-9d4c-159d6c18f53e" (UID: "32de44a4-3f38-456a-9d4c-159d6c18f53e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.913223 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9lz6q\" (UniqueName: \"kubernetes.io/projected/32de44a4-3f38-456a-9d4c-159d6c18f53e-kube-api-access-9lz6q\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.913269 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:22 crc kubenswrapper[4872]: I1009 09:10:22.913284 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/32de44a4-3f38-456a-9d4c-159d6c18f53e-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.369559 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-d6b6m" event={"ID":"32de44a4-3f38-456a-9d4c-159d6c18f53e","Type":"ContainerDied","Data":"453931a4a8231ea34a5513e05c3ac3bbc2f491d57a7835f20fca627d694b32bf"} Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.369612 4872 scope.go:117] "RemoveContainer" containerID="d57dddf97c4c04bdcc70112f3fac34b2f07b0057128c2293241d5a2f30f47bfc" Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.369721 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-d6b6m" Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.401710 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.410630 4872 scope.go:117] "RemoveContainer" containerID="edc2cae936c92eb2b8cd4a07b3c716040a0ae3351e93f1e7cf0871d3cefca4fd" Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.412151 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-d6b6m"] Oct 09 09:10:23 crc kubenswrapper[4872]: I1009 09:10:23.430960 4872 scope.go:117] "RemoveContainer" containerID="e6e6b806e90ef198f82112b2333e94f45fd602cfd0175abf808f7b30751d6d10" Oct 09 09:10:24 crc kubenswrapper[4872]: I1009 09:10:24.472261 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" path="/var/lib/kubelet/pods/32de44a4-3f38-456a-9d4c-159d6c18f53e/volumes" Oct 09 09:10:30 crc kubenswrapper[4872]: I1009 09:10:30.461335 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:10:30 crc kubenswrapper[4872]: E1009 09:10:30.462271 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:10:41 crc kubenswrapper[4872]: I1009 09:10:41.462124 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:10:41 crc kubenswrapper[4872]: E1009 09:10:41.462869 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:10:52 crc kubenswrapper[4872]: I1009 09:10:52.468741 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:10:52 crc kubenswrapper[4872]: E1009 09:10:52.471574 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:11:04 crc kubenswrapper[4872]: I1009 09:11:04.461988 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:11:04 crc kubenswrapper[4872]: E1009 09:11:04.462869 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:11:15 crc kubenswrapper[4872]: I1009 09:11:15.461533 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:11:15 crc kubenswrapper[4872]: E1009 09:11:15.462665 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:11:30 crc kubenswrapper[4872]: I1009 09:11:30.463011 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:11:30 crc kubenswrapper[4872]: E1009 09:11:30.465469 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:11:42 crc kubenswrapper[4872]: I1009 09:11:42.475239 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:11:42 crc kubenswrapper[4872]: E1009 09:11:42.476217 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:11:53 crc kubenswrapper[4872]: I1009 09:11:53.462763 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:11:53 crc kubenswrapper[4872]: E1009 09:11:53.464440 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:12:05 crc kubenswrapper[4872]: I1009 09:12:05.461834 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:12:05 crc kubenswrapper[4872]: E1009 09:12:05.463850 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:12:20 crc kubenswrapper[4872]: I1009 09:12:20.461470 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:12:21 crc kubenswrapper[4872]: I1009 09:12:21.457896 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6"} Oct 09 09:14:38 crc kubenswrapper[4872]: I1009 09:14:38.157255 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:14:38 crc kubenswrapper[4872]: I1009 09:14:38.158141 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:14:48 crc kubenswrapper[4872]: I1009 09:14:48.861966 4872 generic.go:334] "Generic (PLEG): container finished" podID="efed60a8-77a0-4635-8ea2-46f3400b9090" containerID="2510a36c00800f7f7dc1123d164256f3f304093b69293f95ab77e1f85d648116" exitCode=0 Oct 09 09:14:48 crc kubenswrapper[4872]: I1009 09:14:48.862039 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efed60a8-77a0-4635-8ea2-46f3400b9090","Type":"ContainerDied","Data":"2510a36c00800f7f7dc1123d164256f3f304093b69293f95ab77e1f85d648116"} Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.215299 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.262942 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263096 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263179 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263229 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263309 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263353 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263398 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vhpq\" (UniqueName: \"kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263442 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.263499 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config\") pod \"efed60a8-77a0-4635-8ea2-46f3400b9090\" (UID: \"efed60a8-77a0-4635-8ea2-46f3400b9090\") " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.264735 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data" (OuterVolumeSpecName: "config-data") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.264911 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.270467 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "test-operator-logs") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.274218 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.274798 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq" (OuterVolumeSpecName: "kube-api-access-6vhpq") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "kube-api-access-6vhpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.298191 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.301218 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.302303 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.321116 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "efed60a8-77a0-4635-8ea2-46f3400b9090" (UID: "efed60a8-77a0-4635-8ea2-46f3400b9090"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365250 4872 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365292 4872 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ca-certs\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365306 4872 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365321 4872 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-ssh-key\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365336 4872 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/efed60a8-77a0-4635-8ea2-46f3400b9090-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365348 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vhpq\" (UniqueName: \"kubernetes.io/projected/efed60a8-77a0-4635-8ea2-46f3400b9090-kube-api-access-6vhpq\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365360 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365372 4872 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-openstack-config\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.365383 4872 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/efed60a8-77a0-4635-8ea2-46f3400b9090-config-data\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.393590 4872 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.466962 4872 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.881510 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"efed60a8-77a0-4635-8ea2-46f3400b9090","Type":"ContainerDied","Data":"23b1367df2bde98b4de21401a9f86e89aa3052400bfc5b3e59af96463a9ef8e5"} Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.881566 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b1367df2bde98b4de21401a9f86e89aa3052400bfc5b3e59af96463a9ef8e5" Oct 09 09:14:50 crc kubenswrapper[4872]: I1009 09:14:50.881693 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.186291 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7"] Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187298 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="extract-content" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187316 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="extract-content" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187336 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187345 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187379 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efed60a8-77a0-4635-8ea2-46f3400b9090" containerName="tempest-tests-tempest-tests-runner" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187387 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="efed60a8-77a0-4635-8ea2-46f3400b9090" containerName="tempest-tests-tempest-tests-runner" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187404 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="extract-utilities" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187411 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="extract-utilities" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187437 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187445 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187461 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="extract-content" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187469 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="extract-content" Oct 09 09:15:00 crc kubenswrapper[4872]: E1009 09:15:00.187483 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="extract-utilities" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187491 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="extract-utilities" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187758 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d884fe5-01ce-407d-acf8-10ad86c68122" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187780 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="32de44a4-3f38-456a-9d4c-159d6c18f53e" containerName="registry-server" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.187810 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="efed60a8-77a0-4635-8ea2-46f3400b9090" containerName="tempest-tests-tempest-tests-runner" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.188473 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.191242 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.191453 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.195519 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7"] Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.254093 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.254362 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zpmzf\" (UniqueName: \"kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.254422 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.356286 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.356382 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zpmzf\" (UniqueName: \"kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.356406 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.357407 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.363497 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.373421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zpmzf\" (UniqueName: \"kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf\") pod \"collect-profiles-29333355-h9pt7\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.516023 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.941462 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7"] Oct 09 09:15:00 crc kubenswrapper[4872]: I1009 09:15:00.979549 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" event={"ID":"35866c2d-c731-402a-9958-644ee5eb185e","Type":"ContainerStarted","Data":"b5b88dc6eec1afe6d8685d8207fffd8bc69f499868a54f7b3c6b3e2015f1581a"} Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.916957 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.920246 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.922165 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-2fckw" Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.932527 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.990896 4872 generic.go:334] "Generic (PLEG): container finished" podID="35866c2d-c731-402a-9958-644ee5eb185e" containerID="615cc055d65a2a6e62089f6372e6eef65175c5639142a7c0c9ac80fa458f7306" exitCode=0 Oct 09 09:15:01 crc kubenswrapper[4872]: I1009 09:15:01.991013 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" event={"ID":"35866c2d-c731-402a-9958-644ee5eb185e","Type":"ContainerDied","Data":"615cc055d65a2a6e62089f6372e6eef65175c5639142a7c0c9ac80fa458f7306"} Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.089753 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.090172 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtm4s\" (UniqueName: \"kubernetes.io/projected/cd9e08b7-6f45-42d6-be93-5534318c54cd-kube-api-access-dtm4s\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.191982 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtm4s\" (UniqueName: \"kubernetes.io/projected/cd9e08b7-6f45-42d6-be93-5534318c54cd-kube-api-access-dtm4s\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.192049 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.192581 4872 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.209981 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtm4s\" (UniqueName: \"kubernetes.io/projected/cd9e08b7-6f45-42d6-be93-5534318c54cd-kube-api-access-dtm4s\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.216964 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"cd9e08b7-6f45-42d6-be93-5534318c54cd\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.247613 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.678865 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Oct 09 09:15:02 crc kubenswrapper[4872]: I1009 09:15:02.682620 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.001381 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"cd9e08b7-6f45-42d6-be93-5534318c54cd","Type":"ContainerStarted","Data":"001e2a7505a3ee724daca5adc9d816ee95553331448523912dfc8bac8519cb20"} Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.350570 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.520972 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume\") pod \"35866c2d-c731-402a-9958-644ee5eb185e\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.521137 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zpmzf\" (UniqueName: \"kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf\") pod \"35866c2d-c731-402a-9958-644ee5eb185e\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.521709 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume\") pod \"35866c2d-c731-402a-9958-644ee5eb185e\" (UID: \"35866c2d-c731-402a-9958-644ee5eb185e\") " Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.522336 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume" (OuterVolumeSpecName: "config-volume") pod "35866c2d-c731-402a-9958-644ee5eb185e" (UID: "35866c2d-c731-402a-9958-644ee5eb185e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.530303 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "35866c2d-c731-402a-9958-644ee5eb185e" (UID: "35866c2d-c731-402a-9958-644ee5eb185e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.530497 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf" (OuterVolumeSpecName: "kube-api-access-zpmzf") pod "35866c2d-c731-402a-9958-644ee5eb185e" (UID: "35866c2d-c731-402a-9958-644ee5eb185e"). InnerVolumeSpecName "kube-api-access-zpmzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.623120 4872 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/35866c2d-c731-402a-9958-644ee5eb185e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.623156 4872 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/35866c2d-c731-402a-9958-644ee5eb185e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 09 09:15:03 crc kubenswrapper[4872]: I1009 09:15:03.623166 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zpmzf\" (UniqueName: \"kubernetes.io/projected/35866c2d-c731-402a-9958-644ee5eb185e-kube-api-access-zpmzf\") on node \"crc\" DevicePath \"\"" Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.011015 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"cd9e08b7-6f45-42d6-be93-5534318c54cd","Type":"ContainerStarted","Data":"d2bfa18e805ed05d38d25ea0b77e2f1bbcbb13871c12ec31e34836f92a40a596"} Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.012571 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" event={"ID":"35866c2d-c731-402a-9958-644ee5eb185e","Type":"ContainerDied","Data":"b5b88dc6eec1afe6d8685d8207fffd8bc69f499868a54f7b3c6b3e2015f1581a"} Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.012613 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29333355-h9pt7" Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.012665 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5b88dc6eec1afe6d8685d8207fffd8bc69f499868a54f7b3c6b3e2015f1581a" Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.026429 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.9158611859999999 podStartE2EDuration="3.026410762s" podCreationTimestamp="2025-10-09 09:15:01 +0000 UTC" firstStartedPulling="2025-10-09 09:15:02.682377588 +0000 UTC m=+3460.872906214" lastFinishedPulling="2025-10-09 09:15:03.792927164 +0000 UTC m=+3461.983455790" observedRunningTime="2025-10-09 09:15:04.024106095 +0000 UTC m=+3462.214634741" watchObservedRunningTime="2025-10-09 09:15:04.026410762 +0000 UTC m=+3462.216939388" Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.427538 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h"] Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.435855 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29333310-cjj7h"] Oct 09 09:15:04 crc kubenswrapper[4872]: I1009 09:15:04.478730 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1eb69287-4970-49c0-a89b-28372b40743c" path="/var/lib/kubelet/pods/1eb69287-4970-49c0-a89b-28372b40743c/volumes" Oct 09 09:15:08 crc kubenswrapper[4872]: I1009 09:15:08.157586 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:15:08 crc kubenswrapper[4872]: I1009 09:15:08.158029 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.118746 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5zt8/must-gather-cgs8q"] Oct 09 09:15:21 crc kubenswrapper[4872]: E1009 09:15:21.119588 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35866c2d-c731-402a-9958-644ee5eb185e" containerName="collect-profiles" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.119601 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="35866c2d-c731-402a-9958-644ee5eb185e" containerName="collect-profiles" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.119849 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="35866c2d-c731-402a-9958-644ee5eb185e" containerName="collect-profiles" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.120691 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5zt8/must-gather-cgs8q"] Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.120772 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.124543 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5zt8"/"openshift-service-ca.crt" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.124783 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-b5zt8"/"kube-root-ca.crt" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.124935 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-b5zt8"/"default-dockercfg-8qhk8" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.274789 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.274856 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9qzd\" (UniqueName: \"kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.376378 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.376456 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9qzd\" (UniqueName: \"kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.376811 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.395245 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9qzd\" (UniqueName: \"kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd\") pod \"must-gather-cgs8q\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.437661 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:15:21 crc kubenswrapper[4872]: I1009 09:15:21.918596 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-b5zt8/must-gather-cgs8q"] Oct 09 09:15:22 crc kubenswrapper[4872]: I1009 09:15:22.195521 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" event={"ID":"3640aaac-d2f9-4988-8df1-7fc180f024a7","Type":"ContainerStarted","Data":"eb3cce62997421fd7cd2efcc6e952de904943f036cffefde61ed8a3c7e911b7a"} Oct 09 09:15:26 crc kubenswrapper[4872]: I1009 09:15:26.229377 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" event={"ID":"3640aaac-d2f9-4988-8df1-7fc180f024a7","Type":"ContainerStarted","Data":"8b4dadc50f2abab0b5b409a6f2bef258a4417350d00f0af24515eb807dd6f0e6"} Oct 09 09:15:27 crc kubenswrapper[4872]: I1009 09:15:27.240288 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" event={"ID":"3640aaac-d2f9-4988-8df1-7fc180f024a7","Type":"ContainerStarted","Data":"f8207f8db8734c8b6a9dde736937431b3bdd62534aadf2ee7d6a04896deae635"} Oct 09 09:15:27 crc kubenswrapper[4872]: I1009 09:15:27.261265 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" podStartSLOduration=2.264563222 podStartE2EDuration="6.261246178s" podCreationTimestamp="2025-10-09 09:15:21 +0000 UTC" firstStartedPulling="2025-10-09 09:15:21.923337241 +0000 UTC m=+3480.113865877" lastFinishedPulling="2025-10-09 09:15:25.920020207 +0000 UTC m=+3484.110548833" observedRunningTime="2025-10-09 09:15:27.259452916 +0000 UTC m=+3485.449981542" watchObservedRunningTime="2025-10-09 09:15:27.261246178 +0000 UTC m=+3485.451774804" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.767341 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-jnkhd"] Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.769166 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.862749 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnnbl\" (UniqueName: \"kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.863059 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.965153 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnnbl\" (UniqueName: \"kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.965200 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.965291 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:29 crc kubenswrapper[4872]: I1009 09:15:29.984897 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnnbl\" (UniqueName: \"kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl\") pod \"crc-debug-jnkhd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:30 crc kubenswrapper[4872]: I1009 09:15:30.088155 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:15:30 crc kubenswrapper[4872]: W1009 09:15:30.121612 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9a57fad_d30a_4344_9152_8973f053ecfd.slice/crio-5b67b36ca9a560cd127f5c1d32669c3257dc172dd41c6f4e2f742cd36e2e3ff6 WatchSource:0}: Error finding container 5b67b36ca9a560cd127f5c1d32669c3257dc172dd41c6f4e2f742cd36e2e3ff6: Status 404 returned error can't find the container with id 5b67b36ca9a560cd127f5c1d32669c3257dc172dd41c6f4e2f742cd36e2e3ff6 Oct 09 09:15:30 crc kubenswrapper[4872]: I1009 09:15:30.266182 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" event={"ID":"e9a57fad-d30a-4344-9152-8973f053ecfd","Type":"ContainerStarted","Data":"5b67b36ca9a560cd127f5c1d32669c3257dc172dd41c6f4e2f742cd36e2e3ff6"} Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.157273 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.157977 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.158038 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.158829 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.158892 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6" gracePeriod=600 Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.397298 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6" exitCode=0 Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.397390 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6"} Oct 09 09:15:38 crc kubenswrapper[4872]: I1009 09:15:38.397686 4872 scope.go:117] "RemoveContainer" containerID="bb079255a4872d041a73d9900889df8127462b78157f4aecea8fb722fb06edc2" Oct 09 09:15:41 crc kubenswrapper[4872]: I1009 09:15:41.426379 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" event={"ID":"e9a57fad-d30a-4344-9152-8973f053ecfd","Type":"ContainerStarted","Data":"4547c0e26917a5b16f47d51d3743858e8a92d7672c342cc34f20fac1d93811f3"} Oct 09 09:15:41 crc kubenswrapper[4872]: I1009 09:15:41.429618 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94"} Oct 09 09:15:41 crc kubenswrapper[4872]: I1009 09:15:41.444763 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" podStartSLOduration=1.650777437 podStartE2EDuration="12.444740332s" podCreationTimestamp="2025-10-09 09:15:29 +0000 UTC" firstStartedPulling="2025-10-09 09:15:30.122897744 +0000 UTC m=+3488.313426370" lastFinishedPulling="2025-10-09 09:15:40.916860639 +0000 UTC m=+3499.107389265" observedRunningTime="2025-10-09 09:15:41.43678744 +0000 UTC m=+3499.627316076" watchObservedRunningTime="2025-10-09 09:15:41.444740332 +0000 UTC m=+3499.635268958" Oct 09 09:15:48 crc kubenswrapper[4872]: I1009 09:15:48.462484 4872 scope.go:117] "RemoveContainer" containerID="c9cb5738f683c120377f608e0df0a20b4d9b24150d71ec4cfe22a12a724b0fc5" Oct 09 09:16:18 crc kubenswrapper[4872]: I1009 09:16:18.761184 4872 generic.go:334] "Generic (PLEG): container finished" podID="e9a57fad-d30a-4344-9152-8973f053ecfd" containerID="4547c0e26917a5b16f47d51d3743858e8a92d7672c342cc34f20fac1d93811f3" exitCode=0 Oct 09 09:16:18 crc kubenswrapper[4872]: I1009 09:16:18.761275 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" event={"ID":"e9a57fad-d30a-4344-9152-8973f053ecfd","Type":"ContainerDied","Data":"4547c0e26917a5b16f47d51d3743858e8a92d7672c342cc34f20fac1d93811f3"} Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.885152 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.924425 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-jnkhd"] Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.932497 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-jnkhd"] Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.994608 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host\") pod \"e9a57fad-d30a-4344-9152-8973f053ecfd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.994794 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnnbl\" (UniqueName: \"kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl\") pod \"e9a57fad-d30a-4344-9152-8973f053ecfd\" (UID: \"e9a57fad-d30a-4344-9152-8973f053ecfd\") " Oct 09 09:16:19 crc kubenswrapper[4872]: I1009 09:16:19.996664 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host" (OuterVolumeSpecName: "host") pod "e9a57fad-d30a-4344-9152-8973f053ecfd" (UID: "e9a57fad-d30a-4344-9152-8973f053ecfd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.001763 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl" (OuterVolumeSpecName: "kube-api-access-dnnbl") pod "e9a57fad-d30a-4344-9152-8973f053ecfd" (UID: "e9a57fad-d30a-4344-9152-8973f053ecfd"). InnerVolumeSpecName "kube-api-access-dnnbl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.097526 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e9a57fad-d30a-4344-9152-8973f053ecfd-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.097567 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnnbl\" (UniqueName: \"kubernetes.io/projected/e9a57fad-d30a-4344-9152-8973f053ecfd-kube-api-access-dnnbl\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.472786 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9a57fad-d30a-4344-9152-8973f053ecfd" path="/var/lib/kubelet/pods/e9a57fad-d30a-4344-9152-8973f053ecfd/volumes" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.783219 4872 scope.go:117] "RemoveContainer" containerID="4547c0e26917a5b16f47d51d3743858e8a92d7672c342cc34f20fac1d93811f3" Oct 09 09:16:20 crc kubenswrapper[4872]: I1009 09:16:20.783252 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-jnkhd" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.163063 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-5khrh"] Oct 09 09:16:21 crc kubenswrapper[4872]: E1009 09:16:21.164466 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9a57fad-d30a-4344-9152-8973f053ecfd" containerName="container-00" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.164577 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9a57fad-d30a-4344-9152-8973f053ecfd" containerName="container-00" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.164864 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9a57fad-d30a-4344-9152-8973f053ecfd" containerName="container-00" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.165546 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.219504 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f78n2\" (UniqueName: \"kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.219848 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.322060 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f78n2\" (UniqueName: \"kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.322269 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.322393 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.339820 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f78n2\" (UniqueName: \"kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2\") pod \"crc-debug-5khrh\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.486318 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:21 crc kubenswrapper[4872]: I1009 09:16:21.803201 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" event={"ID":"ac52a559-7885-48bf-b081-6fa375f981dc","Type":"ContainerStarted","Data":"5259353e6cbb72444bab62fc1ed00e1cb889d66c9afb8a6fe56b9dfdb43684ee"} Oct 09 09:16:22 crc kubenswrapper[4872]: I1009 09:16:22.817344 4872 generic.go:334] "Generic (PLEG): container finished" podID="ac52a559-7885-48bf-b081-6fa375f981dc" containerID="27dbafcc3500f6c4ed9dc4315584596afcc774100f1691343b2171620f2d425a" exitCode=0 Oct 09 09:16:22 crc kubenswrapper[4872]: I1009 09:16:22.817798 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" event={"ID":"ac52a559-7885-48bf-b081-6fa375f981dc","Type":"ContainerDied","Data":"27dbafcc3500f6c4ed9dc4315584596afcc774100f1691343b2171620f2d425a"} Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.428892 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-5khrh"] Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.436467 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-5khrh"] Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.930922 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.980170 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f78n2\" (UniqueName: \"kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2\") pod \"ac52a559-7885-48bf-b081-6fa375f981dc\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.980286 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host\") pod \"ac52a559-7885-48bf-b081-6fa375f981dc\" (UID: \"ac52a559-7885-48bf-b081-6fa375f981dc\") " Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.980365 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host" (OuterVolumeSpecName: "host") pod "ac52a559-7885-48bf-b081-6fa375f981dc" (UID: "ac52a559-7885-48bf-b081-6fa375f981dc"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.980861 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ac52a559-7885-48bf-b081-6fa375f981dc-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:23 crc kubenswrapper[4872]: I1009 09:16:23.988431 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2" (OuterVolumeSpecName: "kube-api-access-f78n2") pod "ac52a559-7885-48bf-b081-6fa375f981dc" (UID: "ac52a559-7885-48bf-b081-6fa375f981dc"). InnerVolumeSpecName "kube-api-access-f78n2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.084436 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f78n2\" (UniqueName: \"kubernetes.io/projected/ac52a559-7885-48bf-b081-6fa375f981dc-kube-api-access-f78n2\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.472346 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac52a559-7885-48bf-b081-6fa375f981dc" path="/var/lib/kubelet/pods/ac52a559-7885-48bf-b081-6fa375f981dc/volumes" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.586266 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-kgw8k"] Oct 09 09:16:24 crc kubenswrapper[4872]: E1009 09:16:24.586720 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac52a559-7885-48bf-b081-6fa375f981dc" containerName="container-00" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.586742 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac52a559-7885-48bf-b081-6fa375f981dc" containerName="container-00" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.587031 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac52a559-7885-48bf-b081-6fa375f981dc" containerName="container-00" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.587740 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.696119 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bkvg\" (UniqueName: \"kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.696443 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.797977 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.798156 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bkvg\" (UniqueName: \"kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.798419 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.815816 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bkvg\" (UniqueName: \"kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg\") pod \"crc-debug-kgw8k\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.837928 4872 scope.go:117] "RemoveContainer" containerID="27dbafcc3500f6c4ed9dc4315584596afcc774100f1691343b2171620f2d425a" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.837992 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-5khrh" Oct 09 09:16:24 crc kubenswrapper[4872]: I1009 09:16:24.905524 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:25 crc kubenswrapper[4872]: I1009 09:16:25.856776 4872 generic.go:334] "Generic (PLEG): container finished" podID="404b3dd6-041d-42a0-9e2e-432017cc4d15" containerID="e23bce3bd05b6760b0b07f0a9f9aafc70d3eee6ddc7dfc318dcb7254987ec564" exitCode=0 Oct 09 09:16:25 crc kubenswrapper[4872]: I1009 09:16:25.856963 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" event={"ID":"404b3dd6-041d-42a0-9e2e-432017cc4d15","Type":"ContainerDied","Data":"e23bce3bd05b6760b0b07f0a9f9aafc70d3eee6ddc7dfc318dcb7254987ec564"} Oct 09 09:16:25 crc kubenswrapper[4872]: I1009 09:16:25.857169 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" event={"ID":"404b3dd6-041d-42a0-9e2e-432017cc4d15","Type":"ContainerStarted","Data":"9578546c995e0fb92a52bc20a5ebaf6f5ca3bca773a813711b627fcd7192f098"} Oct 09 09:16:25 crc kubenswrapper[4872]: I1009 09:16:25.903093 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-kgw8k"] Oct 09 09:16:25 crc kubenswrapper[4872]: I1009 09:16:25.912341 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5zt8/crc-debug-kgw8k"] Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.350899 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58c4bcb5d6-tl5n8_4fa33440-7187-4d50-8fe8-b7d6af5db927/barbican-api/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.354070 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58c4bcb5d6-tl5n8_4fa33440-7187-4d50-8fe8-b7d6af5db927/barbican-api-log/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.529260 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5b8ccbfb5d-khqh7_d6806fde-b186-4daa-b56d-e746704bf9a7/barbican-keystone-listener/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.558923 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5b8ccbfb5d-khqh7_d6806fde-b186-4daa-b56d-e746704bf9a7/barbican-keystone-listener-log/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.693032 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bb59858c-hh8zx_0126d82a-0297-4a4e-b75a-1e0a8f31d9d4/barbican-worker/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.720735 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bb59858c-hh8zx_0126d82a-0297-4a4e-b75a-1e0a8f31d9d4/barbican-worker-log/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.891903 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx_441768e0-9ee7-4d91-a9e1-d72ef8fb47e3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.986630 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:26 crc kubenswrapper[4872]: I1009 09:16:26.992484 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/ceilometer-central-agent/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.042783 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6bkvg\" (UniqueName: \"kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg\") pod \"404b3dd6-041d-42a0-9e2e-432017cc4d15\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.042864 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host\") pod \"404b3dd6-041d-42a0-9e2e-432017cc4d15\" (UID: \"404b3dd6-041d-42a0-9e2e-432017cc4d15\") " Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.043010 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host" (OuterVolumeSpecName: "host") pod "404b3dd6-041d-42a0-9e2e-432017cc4d15" (UID: "404b3dd6-041d-42a0-9e2e-432017cc4d15"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.043463 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/404b3dd6-041d-42a0-9e2e-432017cc4d15-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.053539 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg" (OuterVolumeSpecName: "kube-api-access-6bkvg") pod "404b3dd6-041d-42a0-9e2e-432017cc4d15" (UID: "404b3dd6-041d-42a0-9e2e-432017cc4d15"). InnerVolumeSpecName "kube-api-access-6bkvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.064602 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/ceilometer-notification-agent/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.122568 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/proxy-httpd/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.145516 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6bkvg\" (UniqueName: \"kubernetes.io/projected/404b3dd6-041d-42a0-9e2e-432017cc4d15-kube-api-access-6bkvg\") on node \"crc\" DevicePath \"\"" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.193118 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/sg-core/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.310631 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10903b59-1315-45e1-b734-40c85cfa17e7/cinder-api/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.331098 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10903b59-1315-45e1-b734-40c85cfa17e7/cinder-api-log/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.446504 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b221b0fe-0514-409e-9ee6-966c6ceeb683/cinder-scheduler/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.502345 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b221b0fe-0514-409e-9ee6-966c6ceeb683/probe/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.638364 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg_57a6c64a-020d-4b90-b350-84f9441d08b3/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.691228 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-f24mq_f0febe64-b028-4782-a3e8-31b07031c230/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.867203 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv_cc650d56-3744-41ca-90b1-707c8fc60bd9/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.878707 4872 scope.go:117] "RemoveContainer" containerID="e23bce3bd05b6760b0b07f0a9f9aafc70d3eee6ddc7dfc318dcb7254987ec564" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.878816 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/crc-debug-kgw8k" Oct 09 09:16:27 crc kubenswrapper[4872]: I1009 09:16:27.957401 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/init/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.092982 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/init/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.194179 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt_9b5aa1fc-253f-4c3b-b3f4-1364130a3a08/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.197562 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/dnsmasq-dns/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.410077 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d651cc86-02b8-4d90-b4a0-3d1a2de1e300/glance-log/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.413845 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d651cc86-02b8-4d90-b4a0-3d1a2de1e300/glance-httpd/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.473841 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="404b3dd6-041d-42a0-9e2e-432017cc4d15" path="/var/lib/kubelet/pods/404b3dd6-041d-42a0-9e2e-432017cc4d15/volumes" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.572323 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a70f38a0-6bd6-45c9-8dc6-6f540842a4be/glance-httpd/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.587436 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a70f38a0-6bd6-45c9-8dc6-6f540842a4be/glance-log/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.726838 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bf66bd496-2sdwc_6d37cb10-16dd-4360-aaf9-b90839973688/horizon/0.log" Oct 09 09:16:28 crc kubenswrapper[4872]: I1009 09:16:28.977014 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n_abc08f25-2678-4241-9f86-d7c6700a7ae6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.092363 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bf66bd496-2sdwc_6d37cb10-16dd-4360-aaf9-b90839973688/horizon-log/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.116925 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qdjb4_eb153425-cc47-4785-be11-49785db1cca7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.318363 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29333341-v6cn6_3c82d19d-ec17-47d9-aace-a14cfb593965/keystone-cron/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.431575 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7489b48876-4zz4g_7f746daf-5432-4790-961e-c02910ec5946/keystone-api/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.515970 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4d540b3a-5230-484c-b807-ce072ce235d9/kube-state-metrics/0.log" Oct 09 09:16:29 crc kubenswrapper[4872]: I1009 09:16:29.677560 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-dvntq_80ca0a43-215b-42e4-8278-c7ded62e8080/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.126488 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c68d5ff89-tcgwk_7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0/neutron-httpd/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.131935 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c68d5ff89-tcgwk_7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0/neutron-api/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.215192 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl_92e98033-47ad-4e23-a3c4-d8c886be6d0c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.705344 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7a95ce78-4b3a-4aea-8351-c39dc3836f4c/nova-api-log/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.787941 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c572e5bd-51c3-49b9-9c55-7e9f888926a2/nova-cell0-conductor-conductor/0.log" Oct 09 09:16:30 crc kubenswrapper[4872]: I1009 09:16:30.903117 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7a95ce78-4b3a-4aea-8351-c39dc3836f4c/nova-api-api/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.039477 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_cf29706b-8cae-42b7-ba76-b52189349378/nova-cell1-conductor-conductor/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.119041 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_15c4d286-8c8e-40a2-b034-77f11abf9a08/nova-cell1-novncproxy-novncproxy/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.290510 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-h2wx7_fc88437f-6937-47eb-bebd-514527660dd2/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.474796 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6f6917d5-6f1e-4f7f-b651-14ee02e1582c/nova-metadata-log/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.706023 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_605ee569-e22f-4c02-8a31-34a9a92d7497/nova-scheduler-scheduler/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.733238 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/mysql-bootstrap/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.928876 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/mysql-bootstrap/0.log" Oct 09 09:16:31 crc kubenswrapper[4872]: I1009 09:16:31.964769 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/galera/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.148994 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/mysql-bootstrap/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.454601 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/mysql-bootstrap/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.458131 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/galera/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.598217 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6f6917d5-6f1e-4f7f-b651-14ee02e1582c/nova-metadata-metadata/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.681145 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_49ba1d52-eb34-49dd-8e46-34219f2e8336/openstackclient/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.774311 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7p7x8_19bf463a-82b1-4522-a8c7-0661281c1e50/openstack-network-exporter/0.log" Oct 09 09:16:32 crc kubenswrapper[4872]: I1009 09:16:32.898297 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server-init/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.146827 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.241308 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovs-vswitchd/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.258452 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server-init/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.365710 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zk2nb_2e454e39-b253-4330-b5f7-7619c2f0f1c9/ovn-controller/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.509212 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-wwvpw_d9ac9cf3-10a8-45ef-b729-4822000bdc63/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.544603 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85/openstack-network-exporter/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.713562 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85/ovn-northd/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.775811 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a/openstack-network-exporter/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.817118 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a/ovsdbserver-nb/0.log" Oct 09 09:16:33 crc kubenswrapper[4872]: I1009 09:16:33.961703 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_81338d5d-c3ed-4163-90c3-0e60e578e35b/openstack-network-exporter/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.043492 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_81338d5d-c3ed-4163-90c3-0e60e578e35b/ovsdbserver-sb/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.250977 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66db9d8d8d-cz7s2_f4ad5d2f-42db-4d64-9d37-d37e89539c20/placement-api/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.334506 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66db9d8d8d-cz7s2_f4ad5d2f-42db-4d64-9d37-d37e89539c20/placement-log/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.423945 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/setup-container/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.771637 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/setup-container/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.777729 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/rabbitmq/0.log" Oct 09 09:16:34 crc kubenswrapper[4872]: I1009 09:16:34.803466 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/setup-container/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.031397 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/setup-container/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.093741 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/rabbitmq/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.166520 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll_710632e7-e59c-4437-a9c5-171a80e989f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.318654 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qd5jg_5bf8bedf-9b23-4570-b728-fc01c39c9373/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.385354 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6_b21bf250-3cfe-47da-8ff1-2c4a54dc7b77/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.574760 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-w99z2_0a5b2694-c0aa-44f4-aec1-4b3dad1a750d/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.653880 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h292d_cefe7874-2185-4237-b9b6-aa8332849f11/ssh-known-hosts-edpm-deployment/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.879528 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b44c8599-9mlwc_dbc66806-aef2-4aa8-86ca-c48f90386f6c/proxy-server/0.log" Oct 09 09:16:35 crc kubenswrapper[4872]: I1009 09:16:35.893250 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b44c8599-9mlwc_dbc66806-aef2-4aa8-86ca-c48f90386f6c/proxy-httpd/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.094169 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qx7q8_31fd23b6-fa18-4681-ba21-b9e69ad958d0/swift-ring-rebalance/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.141087 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-auditor/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.174103 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-reaper/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.320613 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-replicator/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.428300 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-replicator/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.435939 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-auditor/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.456802 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-server/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.554039 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-server/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.654910 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-updater/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.683998 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-auditor/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.688492 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-expirer/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.808445 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-replicator/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.909860 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-server/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.936730 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/rsync/0.log" Oct 09 09:16:36 crc kubenswrapper[4872]: I1009 09:16:36.969572 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-updater/0.log" Oct 09 09:16:37 crc kubenswrapper[4872]: I1009 09:16:37.022820 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/swift-recon-cron/0.log" Oct 09 09:16:37 crc kubenswrapper[4872]: I1009 09:16:37.177340 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-258q8_55becb4c-820e-4d84-b7c3-e3b2bddc84a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:37 crc kubenswrapper[4872]: I1009 09:16:37.290279 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_efed60a8-77a0-4635-8ea2-46f3400b9090/tempest-tests-tempest-tests-runner/0.log" Oct 09 09:16:37 crc kubenswrapper[4872]: I1009 09:16:37.408502 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_cd9e08b7-6f45-42d6-be93-5534318c54cd/test-operator-logs-container/0.log" Oct 09 09:16:37 crc kubenswrapper[4872]: I1009 09:16:37.517297 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-msgrc_3586ab35-c3cc-4f05-8319-b0b568c2c029/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:16:44 crc kubenswrapper[4872]: I1009 09:16:44.224743 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5c8202d4-bb37-4cad-a36c-547b55adf925/memcached/0.log" Oct 09 09:17:00 crc kubenswrapper[4872]: I1009 09:17:00.629071 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:17:00 crc kubenswrapper[4872]: I1009 09:17:00.840950 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:17:00 crc kubenswrapper[4872]: I1009 09:17:00.842507 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:17:00 crc kubenswrapper[4872]: I1009 09:17:00.858330 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.043279 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.059132 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.060066 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/extract/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.209614 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-lrlzc_91db56d0-f7bc-4d02-92ec-c7ead6625176/kube-rbac-proxy/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.272403 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-lrlzc_91db56d0-f7bc-4d02-92ec-c7ead6625176/manager/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.303963 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-vsdt6_283f56cf-99e9-471b-8f40-10ec0f31ad06/kube-rbac-proxy/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.425025 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-vsdt6_283f56cf-99e9-471b-8f40-10ec0f31ad06/manager/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.456054 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-26dhq_0d04d185-02c6-4688-83cd-e9f1744776f1/kube-rbac-proxy/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.489086 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-26dhq_0d04d185-02c6-4688-83cd-e9f1744776f1/manager/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.655954 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-dskmk_fab5d9a4-352a-4092-9139-c74242142eda/kube-rbac-proxy/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.711381 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-dskmk_fab5d9a4-352a-4092-9139-c74242142eda/manager/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.827106 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-qnnp4_b9b5d69d-fb26-4286-a9b0-3b00d042972a/kube-rbac-proxy/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.829872 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-qnnp4_b9b5d69d-fb26-4286-a9b0-3b00d042972a/manager/0.log" Oct 09 09:17:01 crc kubenswrapper[4872]: I1009 09:17:01.880342 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-n5rbg_deb85a9b-515a-4027-a178-c83829fd5a34/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.014182 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-n5rbg_deb85a9b-515a-4027-a178-c83829fd5a34/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.074834 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-kgzkf_57a99d6c-df10-4de9-9434-471ceb56878d/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.220010 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-kgzkf_57a99d6c-df10-4de9-9434-471ceb56878d/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.257050 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vqj98_c5d3f64a-0214-457c-b4c6-f48beabe8669/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.283198 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vqj98_c5d3f64a-0214-457c-b4c6-f48beabe8669/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.424808 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-8rkjv_a34a57a1-2b91-4d65-8336-4a363b33dd45/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.527582 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-8rkjv_a34a57a1-2b91-4d65-8336-4a363b33dd45/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.603941 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-2knwl_18e70210-0a78-4ea0-8d8b-a955cdbec6b8/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.625493 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-2knwl_18e70210-0a78-4ea0-8d8b-a955cdbec6b8/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.722149 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-8ttcw_88a9a65b-d7fe-4471-b408-562f5c98bdbb/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.807343 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-8ttcw_88a9a65b-d7fe-4471-b408-562f5c98bdbb/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.855256 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-kz58p_fcb8eabd-193f-4b03-b011-5e18514cc858/kube-rbac-proxy/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.962257 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-kz58p_fcb8eabd-193f-4b03-b011-5e18514cc858/manager/0.log" Oct 09 09:17:02 crc kubenswrapper[4872]: I1009 09:17:02.986198 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-v67sz_ae4070fd-f49f-403d-a0b5-3285f235a221/kube-rbac-proxy/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.138624 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-v67sz_ae4070fd-f49f-403d-a0b5-3285f235a221/manager/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.199963 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-ffks5_c4fa74e8-0166-4704-8a06-401b53e1cefe/kube-rbac-proxy/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.222846 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-ffks5_c4fa74e8-0166-4704-8a06-401b53e1cefe/manager/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.366598 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc_4099adc7-9a49-4191-bfc3-3f17b15312db/manager/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.371238 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc_4099adc7-9a49-4191-bfc3-3f17b15312db/kube-rbac-proxy/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.489619 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f665bd66d-9lhnp_4dfe7dfe-548f-4ec8-a435-72217082dd3b/kube-rbac-proxy/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.655313 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-557bb754c5-pj5t9_02685dc1-985a-4360-8192-74858bb6062b/kube-rbac-proxy/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.865236 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-557bb754c5-pj5t9_02685dc1-985a-4360-8192-74858bb6062b/operator/0.log" Oct 09 09:17:03 crc kubenswrapper[4872]: I1009 09:17:03.937362 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5jngg_6249ff92-0786-431e-9f18-d48d3b2a6863/registry-server/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.087000 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-xphlj_6c78484d-5fa5-4f18-ace9-a6811363dca7/kube-rbac-proxy/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.231236 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-xphlj_6c78484d-5fa5-4f18-ace9-a6811363dca7/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.329118 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-sbzmh_25798330-695b-4a9d-b271-48b6d025a3f8/kube-rbac-proxy/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.394979 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-sbzmh_25798330-695b-4a9d-b271-48b6d025a3f8/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.467671 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz_add3400b-1fe2-4b5e-a810-7cca5a38ac96/operator/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.618628 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f665bd66d-9lhnp_4dfe7dfe-548f-4ec8-a435-72217082dd3b/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.661035 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-w6jlm_bfaf85e3-55b3-4564-9ffc-3b80e7f32c85/kube-rbac-proxy/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.676407 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-w6jlm_bfaf85e3-55b3-4564-9ffc-3b80e7f32c85/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.756117 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-b9zg2_a9091168-5676-4cd3-9a0d-f76d0c7c88df/kube-rbac-proxy/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.826599 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-b9zg2_a9091168-5676-4cd3-9a0d-f76d0c7c88df/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.850178 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-ppxzv_0cf7b66d-2680-43e9-812b-1a3713eea708/kube-rbac-proxy/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.951202 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-ppxzv_0cf7b66d-2680-43e9-812b-1a3713eea708/manager/0.log" Oct 09 09:17:04 crc kubenswrapper[4872]: I1009 09:17:04.993585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjx2l_3b929d55-693e-419d-ad6e-aa0005f6538a/kube-rbac-proxy/0.log" Oct 09 09:17:05 crc kubenswrapper[4872]: I1009 09:17:05.034871 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjx2l_3b929d55-693e-419d-ad6e-aa0005f6538a/manager/0.log" Oct 09 09:17:19 crc kubenswrapper[4872]: I1009 09:17:19.217550 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-b5qvz_1f314406-7129-4061-922a-7c2fba4019d9/control-plane-machine-set-operator/0.log" Oct 09 09:17:19 crc kubenswrapper[4872]: I1009 09:17:19.412353 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wcb54_a1a5b7f5-7caa-4fe2-afb4-6c7176939e63/machine-api-operator/0.log" Oct 09 09:17:19 crc kubenswrapper[4872]: I1009 09:17:19.412478 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wcb54_a1a5b7f5-7caa-4fe2-afb4-6c7176939e63/kube-rbac-proxy/0.log" Oct 09 09:17:30 crc kubenswrapper[4872]: I1009 09:17:30.349655 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-cq5wd_ba374ef8-6f28-4a47-a948-8cd66cff6132/cert-manager-controller/0.log" Oct 09 09:17:30 crc kubenswrapper[4872]: I1009 09:17:30.496279 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r594q_eb652e7d-6150-46ae-956f-0fcf642d935a/cert-manager-cainjector/0.log" Oct 09 09:17:30 crc kubenswrapper[4872]: I1009 09:17:30.547438 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-k456p_f1f5f354-df1e-49fa-a117-8e99befeaf38/cert-manager-webhook/0.log" Oct 09 09:17:41 crc kubenswrapper[4872]: I1009 09:17:41.871745 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-fz8l7_d9654e2a-2300-47ca-8725-682e7d1cee0a/nmstate-console-plugin/0.log" Oct 09 09:17:42 crc kubenswrapper[4872]: I1009 09:17:42.046476 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lgwck_591e63c4-2cfa-478c-a509-e6ad87de2cb5/nmstate-handler/0.log" Oct 09 09:17:42 crc kubenswrapper[4872]: I1009 09:17:42.097444 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fn24v_40f6895e-a74d-4ef0-bb3b-9024d4b2db5d/kube-rbac-proxy/0.log" Oct 09 09:17:42 crc kubenswrapper[4872]: I1009 09:17:42.116140 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fn24v_40f6895e-a74d-4ef0-bb3b-9024d4b2db5d/nmstate-metrics/0.log" Oct 09 09:17:42 crc kubenswrapper[4872]: I1009 09:17:42.254856 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-nwsgr_2d597183-b39b-40c1-ac08-ad5467522fc6/nmstate-operator/0.log" Oct 09 09:17:42 crc kubenswrapper[4872]: I1009 09:17:42.324001 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-ppr2b_d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b/nmstate-webhook/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.537803 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2hm62_fadbcba0-ee7c-4e6b-8134-b60655294553/kube-rbac-proxy/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.549281 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2hm62_fadbcba0-ee7c-4e6b-8134-b60655294553/controller/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.721947 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.914599 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.957779 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.962205 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:17:55 crc kubenswrapper[4872]: I1009 09:17:55.992585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.187292 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.192376 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.193462 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.216106 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.409560 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/controller/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.414602 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.442727 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.450249 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.625315 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/frr-metrics/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.648234 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/kube-rbac-proxy-frr/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.668117 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/kube-rbac-proxy/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.860093 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/reloader/0.log" Oct 09 09:17:56 crc kubenswrapper[4872]: I1009 09:17:56.923485 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-9zqhg_381ccfd5-05cf-4592-933e-128d714bf0a6/frr-k8s-webhook-server/0.log" Oct 09 09:17:57 crc kubenswrapper[4872]: I1009 09:17:57.167279 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d774594fc-m77qs_75ddc562-e0fb-4c1c-b98d-eb754c97b989/manager/0.log" Oct 09 09:17:57 crc kubenswrapper[4872]: I1009 09:17:57.356750 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f4fb54f85-5jlsq_c49185d6-a699-454b-8e58-88d86a55b694/webhook-server/0.log" Oct 09 09:17:57 crc kubenswrapper[4872]: I1009 09:17:57.432424 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-b52wr_76fe700a-cada-4802-91e5-1bd3522ab565/kube-rbac-proxy/0.log" Oct 09 09:17:57 crc kubenswrapper[4872]: I1009 09:17:57.936424 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-b52wr_76fe700a-cada-4802-91e5-1bd3522ab565/speaker/0.log" Oct 09 09:17:57 crc kubenswrapper[4872]: I1009 09:17:57.984381 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/frr/0.log" Oct 09 09:18:08 crc kubenswrapper[4872]: I1009 09:18:08.157157 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:18:08 crc kubenswrapper[4872]: I1009 09:18:08.157681 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.043765 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.236757 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.252729 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.279602 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.435617 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.438234 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.482068 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/extract/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.606063 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.774571 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.784564 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.787616 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.959788 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:18:09 crc kubenswrapper[4872]: I1009 09:18:09.960105 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.182608 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.296322 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/registry-server/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.335077 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.377224 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.400261 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.510916 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.538798 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:18:10 crc kubenswrapper[4872]: I1009 09:18:10.778200 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.015843 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.026147 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/registry-server/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.035067 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.070453 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.214626 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.232058 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/extract/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.258206 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.418802 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.426268 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-69jm8_6b2d20f2-e34f-4b19-85fb-0e0bfffe849d/marketplace-operator/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.615908 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.616266 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.642656 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.797141 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.801233 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.972469 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/registry-server/0.log" Oct 09 09:18:11 crc kubenswrapper[4872]: I1009 09:18:11.991830 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.149392 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.184892 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.210810 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.355973 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.372108 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:18:12 crc kubenswrapper[4872]: I1009 09:18:12.762930 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/registry-server/0.log" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.430712 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:24 crc kubenswrapper[4872]: E1009 09:18:24.431774 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404b3dd6-041d-42a0-9e2e-432017cc4d15" containerName="container-00" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.431787 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="404b3dd6-041d-42a0-9e2e-432017cc4d15" containerName="container-00" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.432007 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="404b3dd6-041d-42a0-9e2e-432017cc4d15" containerName="container-00" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.439329 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.457498 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.572513 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.572964 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.573098 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg6jv\" (UniqueName: \"kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.674392 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.675619 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg6jv\" (UniqueName: \"kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.675752 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.676174 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.674904 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.699565 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg6jv\" (UniqueName: \"kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv\") pod \"redhat-operators-j64nd\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:24 crc kubenswrapper[4872]: I1009 09:18:24.775215 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:25 crc kubenswrapper[4872]: I1009 09:18:25.220098 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:25 crc kubenswrapper[4872]: I1009 09:18:25.883897 4872 generic.go:334] "Generic (PLEG): container finished" podID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerID="f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3" exitCode=0 Oct 09 09:18:25 crc kubenswrapper[4872]: I1009 09:18:25.884006 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerDied","Data":"f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3"} Oct 09 09:18:25 crc kubenswrapper[4872]: I1009 09:18:25.884181 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerStarted","Data":"d44a697a7f361988597902b5174383d366f2ae187715d8287e00509b5a82d4d9"} Oct 09 09:18:27 crc kubenswrapper[4872]: I1009 09:18:27.915486 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerStarted","Data":"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6"} Oct 09 09:18:28 crc kubenswrapper[4872]: I1009 09:18:28.927245 4872 generic.go:334] "Generic (PLEG): container finished" podID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerID="41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6" exitCode=0 Oct 09 09:18:28 crc kubenswrapper[4872]: I1009 09:18:28.927353 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerDied","Data":"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6"} Oct 09 09:18:29 crc kubenswrapper[4872]: I1009 09:18:29.938800 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerStarted","Data":"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593"} Oct 09 09:18:29 crc kubenswrapper[4872]: I1009 09:18:29.955108 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j64nd" podStartSLOduration=2.312520044 podStartE2EDuration="5.955088849s" podCreationTimestamp="2025-10-09 09:18:24 +0000 UTC" firstStartedPulling="2025-10-09 09:18:25.885867153 +0000 UTC m=+3664.076395779" lastFinishedPulling="2025-10-09 09:18:29.528435938 +0000 UTC m=+3667.718964584" observedRunningTime="2025-10-09 09:18:29.953277484 +0000 UTC m=+3668.143806130" watchObservedRunningTime="2025-10-09 09:18:29.955088849 +0000 UTC m=+3668.145617465" Oct 09 09:18:34 crc kubenswrapper[4872]: I1009 09:18:34.776443 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:34 crc kubenswrapper[4872]: I1009 09:18:34.776996 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:35 crc kubenswrapper[4872]: I1009 09:18:35.831632 4872 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j64nd" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="registry-server" probeResult="failure" output=< Oct 09 09:18:35 crc kubenswrapper[4872]: timeout: failed to connect service ":50051" within 1s Oct 09 09:18:35 crc kubenswrapper[4872]: > Oct 09 09:18:38 crc kubenswrapper[4872]: I1009 09:18:38.156771 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:18:38 crc kubenswrapper[4872]: I1009 09:18:38.156852 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:18:44 crc kubenswrapper[4872]: I1009 09:18:44.827429 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:44 crc kubenswrapper[4872]: I1009 09:18:44.887776 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:45 crc kubenswrapper[4872]: I1009 09:18:45.073713 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.107227 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-j64nd" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="registry-server" containerID="cri-o://1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593" gracePeriod=2 Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.570246 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.761906 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg6jv\" (UniqueName: \"kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv\") pod \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.762517 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities\") pod \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.762826 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content\") pod \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\" (UID: \"14ab4d4e-105e-43ce-916e-b92a0aa41d3c\") " Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.763287 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities" (OuterVolumeSpecName: "utilities") pod "14ab4d4e-105e-43ce-916e-b92a0aa41d3c" (UID: "14ab4d4e-105e-43ce-916e-b92a0aa41d3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.763938 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.785551 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv" (OuterVolumeSpecName: "kube-api-access-kg6jv") pod "14ab4d4e-105e-43ce-916e-b92a0aa41d3c" (UID: "14ab4d4e-105e-43ce-916e-b92a0aa41d3c"). InnerVolumeSpecName "kube-api-access-kg6jv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.846558 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "14ab4d4e-105e-43ce-916e-b92a0aa41d3c" (UID: "14ab4d4e-105e-43ce-916e-b92a0aa41d3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.865625 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kg6jv\" (UniqueName: \"kubernetes.io/projected/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-kube-api-access-kg6jv\") on node \"crc\" DevicePath \"\"" Oct 09 09:18:46 crc kubenswrapper[4872]: I1009 09:18:46.865676 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/14ab4d4e-105e-43ce-916e-b92a0aa41d3c-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.121852 4872 generic.go:334] "Generic (PLEG): container finished" podID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerID="1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593" exitCode=0 Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.121939 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerDied","Data":"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593"} Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.121998 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j64nd" event={"ID":"14ab4d4e-105e-43ce-916e-b92a0aa41d3c","Type":"ContainerDied","Data":"d44a697a7f361988597902b5174383d366f2ae187715d8287e00509b5a82d4d9"} Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.122041 4872 scope.go:117] "RemoveContainer" containerID="1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.122487 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j64nd" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.148483 4872 scope.go:117] "RemoveContainer" containerID="41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.201241 4872 scope.go:117] "RemoveContainer" containerID="f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.205111 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.219864 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-j64nd"] Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.234910 4872 scope.go:117] "RemoveContainer" containerID="1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593" Oct 09 09:18:47 crc kubenswrapper[4872]: E1009 09:18:47.235291 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593\": container with ID starting with 1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593 not found: ID does not exist" containerID="1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.235339 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593"} err="failed to get container status \"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593\": rpc error: code = NotFound desc = could not find container \"1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593\": container with ID starting with 1c607c0b137f3178202f841768e97a8f8c4ee874727e06281a2448833f148593 not found: ID does not exist" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.235361 4872 scope.go:117] "RemoveContainer" containerID="41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6" Oct 09 09:18:47 crc kubenswrapper[4872]: E1009 09:18:47.235871 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6\": container with ID starting with 41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6 not found: ID does not exist" containerID="41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.235926 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6"} err="failed to get container status \"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6\": rpc error: code = NotFound desc = could not find container \"41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6\": container with ID starting with 41ad075c18b0c4ec4dbca66f3ad93fde0484c08a5adf4765e013e960475b0da6 not found: ID does not exist" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.235954 4872 scope.go:117] "RemoveContainer" containerID="f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3" Oct 09 09:18:47 crc kubenswrapper[4872]: E1009 09:18:47.236298 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3\": container with ID starting with f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3 not found: ID does not exist" containerID="f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3" Oct 09 09:18:47 crc kubenswrapper[4872]: I1009 09:18:47.236345 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3"} err="failed to get container status \"f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3\": rpc error: code = NotFound desc = could not find container \"f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3\": container with ID starting with f2c6f712fba19e1855b3c1cbc3268475cfc63eefe21bf6fa51ea119cf40caea3 not found: ID does not exist" Oct 09 09:18:48 crc kubenswrapper[4872]: I1009 09:18:48.477370 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" path="/var/lib/kubelet/pods/14ab4d4e-105e-43ce-916e-b92a0aa41d3c/volumes" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.157059 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.157737 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.157795 4872 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.158730 4872 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94"} pod="openshift-machine-config-operator/machine-config-daemon-98kmz" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.158804 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" containerID="cri-o://252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" gracePeriod=600 Oct 09 09:19:08 crc kubenswrapper[4872]: E1009 09:19:08.314706 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.339048 4872 generic.go:334] "Generic (PLEG): container finished" podID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" exitCode=0 Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.339214 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerDied","Data":"252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94"} Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.339308 4872 scope.go:117] "RemoveContainer" containerID="09a44d1f41246150fac8631c7f3da74844bf3d97d316a5f25fce649d9a24a7d6" Oct 09 09:19:08 crc kubenswrapper[4872]: I1009 09:19:08.341056 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:19:08 crc kubenswrapper[4872]: E1009 09:19:08.345153 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:20 crc kubenswrapper[4872]: I1009 09:19:20.462217 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:19:20 crc kubenswrapper[4872]: E1009 09:19:20.462947 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:31 crc kubenswrapper[4872]: I1009 09:19:31.462457 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:19:31 crc kubenswrapper[4872]: E1009 09:19:31.464119 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:42 crc kubenswrapper[4872]: I1009 09:19:42.468890 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:19:42 crc kubenswrapper[4872]: E1009 09:19:42.470987 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:47 crc kubenswrapper[4872]: I1009 09:19:47.733002 4872 generic.go:334] "Generic (PLEG): container finished" podID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerID="8b4dadc50f2abab0b5b409a6f2bef258a4417350d00f0af24515eb807dd6f0e6" exitCode=0 Oct 09 09:19:47 crc kubenswrapper[4872]: I1009 09:19:47.733090 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" event={"ID":"3640aaac-d2f9-4988-8df1-7fc180f024a7","Type":"ContainerDied","Data":"8b4dadc50f2abab0b5b409a6f2bef258a4417350d00f0af24515eb807dd6f0e6"} Oct 09 09:19:47 crc kubenswrapper[4872]: I1009 09:19:47.734877 4872 scope.go:117] "RemoveContainer" containerID="8b4dadc50f2abab0b5b409a6f2bef258a4417350d00f0af24515eb807dd6f0e6" Oct 09 09:19:48 crc kubenswrapper[4872]: I1009 09:19:48.341971 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5zt8_must-gather-cgs8q_3640aaac-d2f9-4988-8df1-7fc180f024a7/gather/0.log" Oct 09 09:19:53 crc kubenswrapper[4872]: I1009 09:19:53.461561 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:19:53 crc kubenswrapper[4872]: E1009 09:19:53.462319 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:19:56 crc kubenswrapper[4872]: I1009 09:19:56.597619 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-b5zt8/must-gather-cgs8q"] Oct 09 09:19:56 crc kubenswrapper[4872]: I1009 09:19:56.598380 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="copy" containerID="cri-o://f8207f8db8734c8b6a9dde736937431b3bdd62534aadf2ee7d6a04896deae635" gracePeriod=2 Oct 09 09:19:56 crc kubenswrapper[4872]: I1009 09:19:56.605803 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-b5zt8/must-gather-cgs8q"] Oct 09 09:19:56 crc kubenswrapper[4872]: I1009 09:19:56.825855 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5zt8_must-gather-cgs8q_3640aaac-d2f9-4988-8df1-7fc180f024a7/copy/0.log" Oct 09 09:19:56 crc kubenswrapper[4872]: I1009 09:19:56.826478 4872 generic.go:334] "Generic (PLEG): container finished" podID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerID="f8207f8db8734c8b6a9dde736937431b3bdd62534aadf2ee7d6a04896deae635" exitCode=143 Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.133329 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5zt8_must-gather-cgs8q_3640aaac-d2f9-4988-8df1-7fc180f024a7/copy/0.log" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.133697 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.292710 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9qzd\" (UniqueName: \"kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd\") pod \"3640aaac-d2f9-4988-8df1-7fc180f024a7\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.292761 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output\") pod \"3640aaac-d2f9-4988-8df1-7fc180f024a7\" (UID: \"3640aaac-d2f9-4988-8df1-7fc180f024a7\") " Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.305259 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd" (OuterVolumeSpecName: "kube-api-access-p9qzd") pod "3640aaac-d2f9-4988-8df1-7fc180f024a7" (UID: "3640aaac-d2f9-4988-8df1-7fc180f024a7"). InnerVolumeSpecName "kube-api-access-p9qzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.398945 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9qzd\" (UniqueName: \"kubernetes.io/projected/3640aaac-d2f9-4988-8df1-7fc180f024a7-kube-api-access-p9qzd\") on node \"crc\" DevicePath \"\"" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.441408 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "3640aaac-d2f9-4988-8df1-7fc180f024a7" (UID: "3640aaac-d2f9-4988-8df1-7fc180f024a7"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.501856 4872 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/3640aaac-d2f9-4988-8df1-7fc180f024a7-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.837175 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-b5zt8_must-gather-cgs8q_3640aaac-d2f9-4988-8df1-7fc180f024a7/copy/0.log" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.837763 4872 scope.go:117] "RemoveContainer" containerID="f8207f8db8734c8b6a9dde736937431b3bdd62534aadf2ee7d6a04896deae635" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.837805 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-b5zt8/must-gather-cgs8q" Oct 09 09:19:57 crc kubenswrapper[4872]: I1009 09:19:57.856731 4872 scope.go:117] "RemoveContainer" containerID="8b4dadc50f2abab0b5b409a6f2bef258a4417350d00f0af24515eb807dd6f0e6" Oct 09 09:19:58 crc kubenswrapper[4872]: I1009 09:19:58.474048 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" path="/var/lib/kubelet/pods/3640aaac-d2f9-4988-8df1-7fc180f024a7/volumes" Oct 09 09:20:08 crc kubenswrapper[4872]: I1009 09:20:08.462613 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:20:08 crc kubenswrapper[4872]: E1009 09:20:08.463462 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:20:22 crc kubenswrapper[4872]: I1009 09:20:22.471410 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:20:22 crc kubenswrapper[4872]: E1009 09:20:22.472851 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:20:35 crc kubenswrapper[4872]: I1009 09:20:35.461996 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:20:35 crc kubenswrapper[4872]: E1009 09:20:35.462867 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.099249 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-84zmz/must-gather-cdxhx"] Oct 09 09:20:46 crc kubenswrapper[4872]: E1009 09:20:46.100187 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="extract-utilities" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100201 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="extract-utilities" Oct 09 09:20:46 crc kubenswrapper[4872]: E1009 09:20:46.100227 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="extract-content" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100233 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="extract-content" Oct 09 09:20:46 crc kubenswrapper[4872]: E1009 09:20:46.100253 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="gather" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100266 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="gather" Oct 09 09:20:46 crc kubenswrapper[4872]: E1009 09:20:46.100282 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="registry-server" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100291 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="registry-server" Oct 09 09:20:46 crc kubenswrapper[4872]: E1009 09:20:46.100309 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="copy" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100315 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="copy" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100527 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="copy" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100553 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ab4d4e-105e-43ce-916e-b92a0aa41d3c" containerName="registry-server" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.100566 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="3640aaac-d2f9-4988-8df1-7fc180f024a7" containerName="gather" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.110877 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.116081 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-84zmz"/"openshift-service-ca.crt" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.116380 4872 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-84zmz"/"default-dockercfg-rmc8r" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.116557 4872 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-84zmz"/"kube-root-ca.crt" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.142034 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-84zmz/must-gather-cdxhx"] Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.257308 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v52xv\" (UniqueName: \"kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.257415 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.359238 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v52xv\" (UniqueName: \"kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.359311 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.359813 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.385734 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v52xv\" (UniqueName: \"kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv\") pod \"must-gather-cdxhx\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.433121 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:20:46 crc kubenswrapper[4872]: I1009 09:20:46.913317 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-84zmz/must-gather-cdxhx"] Oct 09 09:20:47 crc kubenswrapper[4872]: I1009 09:20:47.327883 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/must-gather-cdxhx" event={"ID":"c8f8a617-7e86-49df-be7f-ccf469601c9a","Type":"ContainerStarted","Data":"49d405d791525b46078eefea6e5dc82a657ff9a87abdc7f9066bed4de05542b2"} Oct 09 09:20:47 crc kubenswrapper[4872]: I1009 09:20:47.328169 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/must-gather-cdxhx" event={"ID":"c8f8a617-7e86-49df-be7f-ccf469601c9a","Type":"ContainerStarted","Data":"bf194005e45680cf103278a5e55257a419d7e0a935ed4590e6b6b74dc1b50b32"} Oct 09 09:20:48 crc kubenswrapper[4872]: I1009 09:20:48.337476 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/must-gather-cdxhx" event={"ID":"c8f8a617-7e86-49df-be7f-ccf469601c9a","Type":"ContainerStarted","Data":"2df4b4d4e29304492834013ec6e632691ccfca48b85a131ad22e33ea22cf91d8"} Oct 09 09:20:48 crc kubenswrapper[4872]: I1009 09:20:48.360621 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-84zmz/must-gather-cdxhx" podStartSLOduration=2.360600114 podStartE2EDuration="2.360600114s" podCreationTimestamp="2025-10-09 09:20:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 09:20:48.353809346 +0000 UTC m=+3806.544337982" watchObservedRunningTime="2025-10-09 09:20:48.360600114 +0000 UTC m=+3806.551128740" Oct 09 09:20:48 crc kubenswrapper[4872]: I1009 09:20:48.462622 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:20:48 crc kubenswrapper[4872]: E1009 09:20:48.462866 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:20:50 crc kubenswrapper[4872]: I1009 09:20:50.995242 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-84zmz/crc-debug-dhljh"] Oct 09 09:20:50 crc kubenswrapper[4872]: I1009 09:20:50.997309 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.143766 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.144417 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znxpm\" (UniqueName: \"kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.246758 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znxpm\" (UniqueName: \"kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.246907 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.247031 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.277326 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znxpm\" (UniqueName: \"kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm\") pod \"crc-debug-dhljh\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: I1009 09:20:51.321549 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:20:51 crc kubenswrapper[4872]: W1009 09:20:51.357982 4872 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fed6492_7901_46da_9dc3_191c1b1fd0d7.slice/crio-e90f5ee38f420ee6197b46b1c90fbd38de4b9ddf0f57c15c1126cf8058fe64d7 WatchSource:0}: Error finding container e90f5ee38f420ee6197b46b1c90fbd38de4b9ddf0f57c15c1126cf8058fe64d7: Status 404 returned error can't find the container with id e90f5ee38f420ee6197b46b1c90fbd38de4b9ddf0f57c15c1126cf8058fe64d7 Oct 09 09:20:52 crc kubenswrapper[4872]: I1009 09:20:52.379931 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-dhljh" event={"ID":"7fed6492-7901-46da-9dc3-191c1b1fd0d7","Type":"ContainerStarted","Data":"00dea041a001d8563af6cbfc01ba06d8d5090ce24adec4dda377117e85bcf7d4"} Oct 09 09:20:52 crc kubenswrapper[4872]: I1009 09:20:52.380519 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-dhljh" event={"ID":"7fed6492-7901-46da-9dc3-191c1b1fd0d7","Type":"ContainerStarted","Data":"e90f5ee38f420ee6197b46b1c90fbd38de4b9ddf0f57c15c1126cf8058fe64d7"} Oct 09 09:20:52 crc kubenswrapper[4872]: I1009 09:20:52.394819 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-84zmz/crc-debug-dhljh" podStartSLOduration=2.394794061 podStartE2EDuration="2.394794061s" podCreationTimestamp="2025-10-09 09:20:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-09 09:20:52.392987356 +0000 UTC m=+3810.583515982" watchObservedRunningTime="2025-10-09 09:20:52.394794061 +0000 UTC m=+3810.585322707" Oct 09 09:21:01 crc kubenswrapper[4872]: I1009 09:21:01.465706 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:21:01 crc kubenswrapper[4872]: E1009 09:21:01.466751 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:21:05 crc kubenswrapper[4872]: I1009 09:21:05.967168 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:05 crc kubenswrapper[4872]: I1009 09:21:05.971723 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:05 crc kubenswrapper[4872]: I1009 09:21:05.984532 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.128518 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.128820 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cz99c\" (UniqueName: \"kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.128864 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.230398 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.230459 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cz99c\" (UniqueName: \"kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.230484 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.230942 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.230958 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.267337 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cz99c\" (UniqueName: \"kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c\") pod \"community-operators-nkfck\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.302174 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:06 crc kubenswrapper[4872]: I1009 09:21:06.909451 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:07 crc kubenswrapper[4872]: I1009 09:21:07.513622 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a199be9-6831-4535-91ad-6ade426054e2" containerID="6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861" exitCode=0 Oct 09 09:21:07 crc kubenswrapper[4872]: I1009 09:21:07.513717 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerDied","Data":"6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861"} Oct 09 09:21:07 crc kubenswrapper[4872]: I1009 09:21:07.513950 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerStarted","Data":"32073fa69822b3ebf9be7d72666b955f3ab7d9b3c46e4bddcf4281dd69695ec3"} Oct 09 09:21:07 crc kubenswrapper[4872]: I1009 09:21:07.516906 4872 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 09 09:21:08 crc kubenswrapper[4872]: I1009 09:21:08.532784 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerStarted","Data":"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381"} Oct 09 09:21:09 crc kubenswrapper[4872]: I1009 09:21:09.543941 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a199be9-6831-4535-91ad-6ade426054e2" containerID="9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381" exitCode=0 Oct 09 09:21:09 crc kubenswrapper[4872]: I1009 09:21:09.544046 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerDied","Data":"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381"} Oct 09 09:21:11 crc kubenswrapper[4872]: I1009 09:21:11.596548 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerStarted","Data":"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244"} Oct 09 09:21:11 crc kubenswrapper[4872]: I1009 09:21:11.615733 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nkfck" podStartSLOduration=3.691578514 podStartE2EDuration="6.615704825s" podCreationTimestamp="2025-10-09 09:21:05 +0000 UTC" firstStartedPulling="2025-10-09 09:21:07.516504673 +0000 UTC m=+3825.707033299" lastFinishedPulling="2025-10-09 09:21:10.440630984 +0000 UTC m=+3828.631159610" observedRunningTime="2025-10-09 09:21:11.611442395 +0000 UTC m=+3829.801971031" watchObservedRunningTime="2025-10-09 09:21:11.615704825 +0000 UTC m=+3829.806233511" Oct 09 09:21:15 crc kubenswrapper[4872]: I1009 09:21:15.462047 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:21:15 crc kubenswrapper[4872]: E1009 09:21:15.462661 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.303363 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.303790 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.354094 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.627719 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.630388 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.635250 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.723874 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.725289 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qltsp\" (UniqueName: \"kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.725354 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.725378 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.826787 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.827057 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.827332 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qltsp\" (UniqueName: \"kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.827421 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.827361 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.858179 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qltsp\" (UniqueName: \"kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp\") pod \"certified-operators-xql92\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:16 crc kubenswrapper[4872]: I1009 09:21:16.958836 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:17 crc kubenswrapper[4872]: I1009 09:21:17.447229 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:17 crc kubenswrapper[4872]: I1009 09:21:17.668300 4872 generic.go:334] "Generic (PLEG): container finished" podID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerID="37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504" exitCode=0 Oct 09 09:21:17 crc kubenswrapper[4872]: I1009 09:21:17.668437 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerDied","Data":"37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504"} Oct 09 09:21:17 crc kubenswrapper[4872]: I1009 09:21:17.668485 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerStarted","Data":"50ee6a3f98d4d86ce94542ba69c976299893c156b27246b21ad8535c2fc42764"} Oct 09 09:21:18 crc kubenswrapper[4872]: I1009 09:21:18.997226 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:19 crc kubenswrapper[4872]: I1009 09:21:19.685487 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerStarted","Data":"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0"} Oct 09 09:21:19 crc kubenswrapper[4872]: I1009 09:21:19.685593 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nkfck" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="registry-server" containerID="cri-o://1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244" gracePeriod=2 Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.204232 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.304887 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content\") pod \"2a199be9-6831-4535-91ad-6ade426054e2\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.305116 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities\") pod \"2a199be9-6831-4535-91ad-6ade426054e2\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.305191 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cz99c\" (UniqueName: \"kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c\") pod \"2a199be9-6831-4535-91ad-6ade426054e2\" (UID: \"2a199be9-6831-4535-91ad-6ade426054e2\") " Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.305895 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities" (OuterVolumeSpecName: "utilities") pod "2a199be9-6831-4535-91ad-6ade426054e2" (UID: "2a199be9-6831-4535-91ad-6ade426054e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.313026 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c" (OuterVolumeSpecName: "kube-api-access-cz99c") pod "2a199be9-6831-4535-91ad-6ade426054e2" (UID: "2a199be9-6831-4535-91ad-6ade426054e2"). InnerVolumeSpecName "kube-api-access-cz99c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.372032 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a199be9-6831-4535-91ad-6ade426054e2" (UID: "2a199be9-6831-4535-91ad-6ade426054e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.407199 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cz99c\" (UniqueName: \"kubernetes.io/projected/2a199be9-6831-4535-91ad-6ade426054e2-kube-api-access-cz99c\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.407234 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.407244 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a199be9-6831-4535-91ad-6ade426054e2-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.694815 4872 generic.go:334] "Generic (PLEG): container finished" podID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerID="3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0" exitCode=0 Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.694969 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerDied","Data":"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0"} Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.697927 4872 generic.go:334] "Generic (PLEG): container finished" podID="2a199be9-6831-4535-91ad-6ade426054e2" containerID="1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244" exitCode=0 Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.697949 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerDied","Data":"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244"} Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.697968 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkfck" event={"ID":"2a199be9-6831-4535-91ad-6ade426054e2","Type":"ContainerDied","Data":"32073fa69822b3ebf9be7d72666b955f3ab7d9b3c46e4bddcf4281dd69695ec3"} Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.697987 4872 scope.go:117] "RemoveContainer" containerID="1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.698105 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkfck" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.732381 4872 scope.go:117] "RemoveContainer" containerID="9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.742302 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.749730 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nkfck"] Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.757072 4872 scope.go:117] "RemoveContainer" containerID="6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.817794 4872 scope.go:117] "RemoveContainer" containerID="1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244" Oct 09 09:21:20 crc kubenswrapper[4872]: E1009 09:21:20.818247 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244\": container with ID starting with 1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244 not found: ID does not exist" containerID="1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.818277 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244"} err="failed to get container status \"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244\": rpc error: code = NotFound desc = could not find container \"1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244\": container with ID starting with 1fb9a45c31836ac630bef789eab4ef644cf94639238cb866a13aa2fe0817e244 not found: ID does not exist" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.818296 4872 scope.go:117] "RemoveContainer" containerID="9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381" Oct 09 09:21:20 crc kubenswrapper[4872]: E1009 09:21:20.818596 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381\": container with ID starting with 9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381 not found: ID does not exist" containerID="9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.818614 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381"} err="failed to get container status \"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381\": rpc error: code = NotFound desc = could not find container \"9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381\": container with ID starting with 9e02f75d864248b845339f20b650373c749c4c1034715dd266c1a64543e92381 not found: ID does not exist" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.818630 4872 scope.go:117] "RemoveContainer" containerID="6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861" Oct 09 09:21:20 crc kubenswrapper[4872]: E1009 09:21:20.818964 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861\": container with ID starting with 6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861 not found: ID does not exist" containerID="6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861" Oct 09 09:21:20 crc kubenswrapper[4872]: I1009 09:21:20.818982 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861"} err="failed to get container status \"6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861\": rpc error: code = NotFound desc = could not find container \"6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861\": container with ID starting with 6c78321f2cd2036b939fc3f087e1a747800eb9ce8328fd6dee202e7aed81a861 not found: ID does not exist" Oct 09 09:21:21 crc kubenswrapper[4872]: I1009 09:21:21.707889 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerStarted","Data":"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850"} Oct 09 09:21:21 crc kubenswrapper[4872]: I1009 09:21:21.730891 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xql92" podStartSLOduration=2.307500789 podStartE2EDuration="5.730873655s" podCreationTimestamp="2025-10-09 09:21:16 +0000 UTC" firstStartedPulling="2025-10-09 09:21:17.669960843 +0000 UTC m=+3835.860489469" lastFinishedPulling="2025-10-09 09:21:21.093333709 +0000 UTC m=+3839.283862335" observedRunningTime="2025-10-09 09:21:21.730183564 +0000 UTC m=+3839.920712200" watchObservedRunningTime="2025-10-09 09:21:21.730873655 +0000 UTC m=+3839.921402271" Oct 09 09:21:22 crc kubenswrapper[4872]: I1009 09:21:22.475690 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a199be9-6831-4535-91ad-6ade426054e2" path="/var/lib/kubelet/pods/2a199be9-6831-4535-91ad-6ade426054e2/volumes" Oct 09 09:21:25 crc kubenswrapper[4872]: I1009 09:21:25.756901 4872 generic.go:334] "Generic (PLEG): container finished" podID="7fed6492-7901-46da-9dc3-191c1b1fd0d7" containerID="00dea041a001d8563af6cbfc01ba06d8d5090ce24adec4dda377117e85bcf7d4" exitCode=0 Oct 09 09:21:25 crc kubenswrapper[4872]: I1009 09:21:25.756992 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-dhljh" event={"ID":"7fed6492-7901-46da-9dc3-191c1b1fd0d7","Type":"ContainerDied","Data":"00dea041a001d8563af6cbfc01ba06d8d5090ce24adec4dda377117e85bcf7d4"} Oct 09 09:21:26 crc kubenswrapper[4872]: I1009 09:21:26.892011 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:21:26 crc kubenswrapper[4872]: I1009 09:21:26.923079 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-dhljh"] Oct 09 09:21:26 crc kubenswrapper[4872]: I1009 09:21:26.931379 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-dhljh"] Oct 09 09:21:26 crc kubenswrapper[4872]: I1009 09:21:26.959836 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:26 crc kubenswrapper[4872]: I1009 09:21:26.959882 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.006588 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.025121 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znxpm\" (UniqueName: \"kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm\") pod \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.025615 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host\") pod \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\" (UID: \"7fed6492-7901-46da-9dc3-191c1b1fd0d7\") " Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.025756 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host" (OuterVolumeSpecName: "host") pod "7fed6492-7901-46da-9dc3-191c1b1fd0d7" (UID: "7fed6492-7901-46da-9dc3-191c1b1fd0d7"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.026437 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7fed6492-7901-46da-9dc3-191c1b1fd0d7-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.053360 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm" (OuterVolumeSpecName: "kube-api-access-znxpm") pod "7fed6492-7901-46da-9dc3-191c1b1fd0d7" (UID: "7fed6492-7901-46da-9dc3-191c1b1fd0d7"). InnerVolumeSpecName "kube-api-access-znxpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.128015 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znxpm\" (UniqueName: \"kubernetes.io/projected/7fed6492-7901-46da-9dc3-191c1b1fd0d7-kube-api-access-znxpm\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.799756 4872 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e90f5ee38f420ee6197b46b1c90fbd38de4b9ddf0f57c15c1126cf8058fe64d7" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.799800 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-dhljh" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.854805 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:27 crc kubenswrapper[4872]: I1009 09:21:27.905543 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.102452 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-84zmz/crc-debug-jkvlb"] Oct 09 09:21:28 crc kubenswrapper[4872]: E1009 09:21:28.103121 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="extract-content" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.103204 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="extract-content" Oct 09 09:21:28 crc kubenswrapper[4872]: E1009 09:21:28.103268 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="registry-server" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.103329 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="registry-server" Oct 09 09:21:28 crc kubenswrapper[4872]: E1009 09:21:28.103423 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fed6492-7901-46da-9dc3-191c1b1fd0d7" containerName="container-00" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.103481 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fed6492-7901-46da-9dc3-191c1b1fd0d7" containerName="container-00" Oct 09 09:21:28 crc kubenswrapper[4872]: E1009 09:21:28.103544 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="extract-utilities" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.103609 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="extract-utilities" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.103908 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a199be9-6831-4535-91ad-6ade426054e2" containerName="registry-server" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.104001 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fed6492-7901-46da-9dc3-191c1b1fd0d7" containerName="container-00" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.104773 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.248796 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.248896 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj8tr\" (UniqueName: \"kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.350319 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.350365 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj8tr\" (UniqueName: \"kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.350458 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.367296 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj8tr\" (UniqueName: \"kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr\") pod \"crc-debug-jkvlb\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.425147 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.475773 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fed6492-7901-46da-9dc3-191c1b1fd0d7" path="/var/lib/kubelet/pods/7fed6492-7901-46da-9dc3-191c1b1fd0d7/volumes" Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.811967 4872 generic.go:334] "Generic (PLEG): container finished" podID="d8993bdb-e8b1-403d-b8df-0fa65c9038d0" containerID="661d0c8f26956489f8bfb64ef90e56f8ae26e4d1a51ff69942668494c4dde0d3" exitCode=0 Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.812302 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" event={"ID":"d8993bdb-e8b1-403d-b8df-0fa65c9038d0","Type":"ContainerDied","Data":"661d0c8f26956489f8bfb64ef90e56f8ae26e4d1a51ff69942668494c4dde0d3"} Oct 09 09:21:28 crc kubenswrapper[4872]: I1009 09:21:28.812367 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" event={"ID":"d8993bdb-e8b1-403d-b8df-0fa65c9038d0","Type":"ContainerStarted","Data":"4a031c08d2218875699e5b2bb82b58c94b86a0f6dde01b6eaefce16622a54fa9"} Oct 09 09:21:29 crc kubenswrapper[4872]: I1009 09:21:29.234246 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-jkvlb"] Oct 09 09:21:29 crc kubenswrapper[4872]: I1009 09:21:29.241936 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-jkvlb"] Oct 09 09:21:29 crc kubenswrapper[4872]: I1009 09:21:29.820938 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xql92" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="registry-server" containerID="cri-o://5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850" gracePeriod=2 Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.032784 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.190273 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host\") pod \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.190438 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj8tr\" (UniqueName: \"kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr\") pod \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\" (UID: \"d8993bdb-e8b1-403d-b8df-0fa65c9038d0\") " Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.191052 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host" (OuterVolumeSpecName: "host") pod "d8993bdb-e8b1-403d-b8df-0fa65c9038d0" (UID: "d8993bdb-e8b1-403d-b8df-0fa65c9038d0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.196905 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr" (OuterVolumeSpecName: "kube-api-access-tj8tr") pod "d8993bdb-e8b1-403d-b8df-0fa65c9038d0" (UID: "d8993bdb-e8b1-403d-b8df-0fa65c9038d0"). InnerVolumeSpecName "kube-api-access-tj8tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.273749 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.293297 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj8tr\" (UniqueName: \"kubernetes.io/projected/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-kube-api-access-tj8tr\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.293331 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d8993bdb-e8b1-403d-b8df-0fa65c9038d0-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.394746 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities\") pod \"82c3e6ba-3471-40f2-9062-99a3c4963530\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.394833 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qltsp\" (UniqueName: \"kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp\") pod \"82c3e6ba-3471-40f2-9062-99a3c4963530\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.394922 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content\") pod \"82c3e6ba-3471-40f2-9062-99a3c4963530\" (UID: \"82c3e6ba-3471-40f2-9062-99a3c4963530\") " Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.399281 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities" (OuterVolumeSpecName: "utilities") pod "82c3e6ba-3471-40f2-9062-99a3c4963530" (UID: "82c3e6ba-3471-40f2-9062-99a3c4963530"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.402051 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp" (OuterVolumeSpecName: "kube-api-access-qltsp") pod "82c3e6ba-3471-40f2-9062-99a3c4963530" (UID: "82c3e6ba-3471-40f2-9062-99a3c4963530"). InnerVolumeSpecName "kube-api-access-qltsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.404739 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-84zmz/crc-debug-xw8tw"] Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.405112 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="extract-content" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405128 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="extract-content" Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.405162 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8993bdb-e8b1-403d-b8df-0fa65c9038d0" containerName="container-00" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405171 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8993bdb-e8b1-403d-b8df-0fa65c9038d0" containerName="container-00" Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.405199 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="registry-server" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405208 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="registry-server" Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.405217 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="extract-utilities" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405225 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="extract-utilities" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405460 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerName="registry-server" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.405492 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8993bdb-e8b1-403d-b8df-0fa65c9038d0" containerName="container-00" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.406493 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.453795 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82c3e6ba-3471-40f2-9062-99a3c4963530" (UID: "82c3e6ba-3471-40f2-9062-99a3c4963530"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.462529 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.462817 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.473458 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8993bdb-e8b1-403d-b8df-0fa65c9038d0" path="/var/lib/kubelet/pods/d8993bdb-e8b1-403d-b8df-0fa65c9038d0/volumes" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.497022 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.497083 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pb5c\" (UniqueName: \"kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.497141 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.497153 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qltsp\" (UniqueName: \"kubernetes.io/projected/82c3e6ba-3471-40f2-9062-99a3c4963530-kube-api-access-qltsp\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.497162 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82c3e6ba-3471-40f2-9062-99a3c4963530-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.599565 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.599654 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.600201 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pb5c\" (UniqueName: \"kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.619123 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pb5c\" (UniqueName: \"kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c\") pod \"crc-debug-xw8tw\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.725637 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.835195 4872 generic.go:334] "Generic (PLEG): container finished" podID="82c3e6ba-3471-40f2-9062-99a3c4963530" containerID="5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850" exitCode=0 Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.835671 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xql92" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.836201 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerDied","Data":"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850"} Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.836248 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xql92" event={"ID":"82c3e6ba-3471-40f2-9062-99a3c4963530","Type":"ContainerDied","Data":"50ee6a3f98d4d86ce94542ba69c976299893c156b27246b21ad8535c2fc42764"} Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.836268 4872 scope.go:117] "RemoveContainer" containerID="5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.837123 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" event={"ID":"a0cac16d-6629-4d42-a83b-305bd1356f48","Type":"ContainerStarted","Data":"40eabe811c9485a85391e5f4b78651b542d729e4e4cafdb3ef26179246be3f32"} Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.840444 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-jkvlb" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.922516 4872 scope.go:117] "RemoveContainer" containerID="3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.944316 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.944751 4872 scope.go:117] "RemoveContainer" containerID="37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.961889 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xql92"] Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.998821 4872 scope.go:117] "RemoveContainer" containerID="5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850" Oct 09 09:21:30 crc kubenswrapper[4872]: E1009 09:21:30.999553 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850\": container with ID starting with 5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850 not found: ID does not exist" containerID="5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.999706 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850"} err="failed to get container status \"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850\": rpc error: code = NotFound desc = could not find container \"5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850\": container with ID starting with 5c888686f57f239018024923dd29949e288e98c3835ca863d616bea8574de850 not found: ID does not exist" Oct 09 09:21:30 crc kubenswrapper[4872]: I1009 09:21:30.999847 4872 scope.go:117] "RemoveContainer" containerID="3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0" Oct 09 09:21:31 crc kubenswrapper[4872]: E1009 09:21:31.000312 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0\": container with ID starting with 3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0 not found: ID does not exist" containerID="3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0" Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.000381 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0"} err="failed to get container status \"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0\": rpc error: code = NotFound desc = could not find container \"3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0\": container with ID starting with 3786ac8d6f6511419a3a522a8eaf66efb0267ce028843270673777b4a9ac03d0 not found: ID does not exist" Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.000428 4872 scope.go:117] "RemoveContainer" containerID="37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504" Oct 09 09:21:31 crc kubenswrapper[4872]: E1009 09:21:31.000806 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504\": container with ID starting with 37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504 not found: ID does not exist" containerID="37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504" Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.000925 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504"} err="failed to get container status \"37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504\": rpc error: code = NotFound desc = could not find container \"37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504\": container with ID starting with 37d9bd5a07a1ed43c3674b19bd9002fe062770b9b7002fe687caeb47bd2f1504 not found: ID does not exist" Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.001008 4872 scope.go:117] "RemoveContainer" containerID="661d0c8f26956489f8bfb64ef90e56f8ae26e4d1a51ff69942668494c4dde0d3" Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.851585 4872 generic.go:334] "Generic (PLEG): container finished" podID="a0cac16d-6629-4d42-a83b-305bd1356f48" containerID="810f108bf4bd2ecd510a6a8b6c269acef95f216c480b501d0c1eb44f6d1dcd38" exitCode=0 Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.851693 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" event={"ID":"a0cac16d-6629-4d42-a83b-305bd1356f48","Type":"ContainerDied","Data":"810f108bf4bd2ecd510a6a8b6c269acef95f216c480b501d0c1eb44f6d1dcd38"} Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.881792 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-xw8tw"] Oct 09 09:21:31 crc kubenswrapper[4872]: I1009 09:21:31.892735 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-84zmz/crc-debug-xw8tw"] Oct 09 09:21:32 crc kubenswrapper[4872]: I1009 09:21:32.472114 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82c3e6ba-3471-40f2-9062-99a3c4963530" path="/var/lib/kubelet/pods/82c3e6ba-3471-40f2-9062-99a3c4963530/volumes" Oct 09 09:21:32 crc kubenswrapper[4872]: I1009 09:21:32.971963 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.047046 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pb5c\" (UniqueName: \"kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c\") pod \"a0cac16d-6629-4d42-a83b-305bd1356f48\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.047184 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host\") pod \"a0cac16d-6629-4d42-a83b-305bd1356f48\" (UID: \"a0cac16d-6629-4d42-a83b-305bd1356f48\") " Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.047315 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host" (OuterVolumeSpecName: "host") pod "a0cac16d-6629-4d42-a83b-305bd1356f48" (UID: "a0cac16d-6629-4d42-a83b-305bd1356f48"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.047713 4872 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a0cac16d-6629-4d42-a83b-305bd1356f48-host\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.071125 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c" (OuterVolumeSpecName: "kube-api-access-5pb5c") pod "a0cac16d-6629-4d42-a83b-305bd1356f48" (UID: "a0cac16d-6629-4d42-a83b-305bd1356f48"). InnerVolumeSpecName "kube-api-access-5pb5c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.148955 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pb5c\" (UniqueName: \"kubernetes.io/projected/a0cac16d-6629-4d42-a83b-305bd1356f48-kube-api-access-5pb5c\") on node \"crc\" DevicePath \"\"" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.871830 4872 scope.go:117] "RemoveContainer" containerID="810f108bf4bd2ecd510a6a8b6c269acef95f216c480b501d0c1eb44f6d1dcd38" Oct 09 09:21:33 crc kubenswrapper[4872]: I1009 09:21:33.871923 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/crc-debug-xw8tw" Oct 09 09:21:34 crc kubenswrapper[4872]: I1009 09:21:34.473470 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0cac16d-6629-4d42-a83b-305bd1356f48" path="/var/lib/kubelet/pods/a0cac16d-6629-4d42-a83b-305bd1356f48/volumes" Oct 09 09:21:41 crc kubenswrapper[4872]: I1009 09:21:41.461918 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:21:41 crc kubenswrapper[4872]: E1009 09:21:41.462807 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:21:46 crc kubenswrapper[4872]: I1009 09:21:46.690131 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58c4bcb5d6-tl5n8_4fa33440-7187-4d50-8fe8-b7d6af5db927/barbican-api/0.log" Oct 09 09:21:46 crc kubenswrapper[4872]: I1009 09:21:46.920972 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-58c4bcb5d6-tl5n8_4fa33440-7187-4d50-8fe8-b7d6af5db927/barbican-api-log/0.log" Oct 09 09:21:46 crc kubenswrapper[4872]: I1009 09:21:46.980176 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5b8ccbfb5d-khqh7_d6806fde-b186-4daa-b56d-e746704bf9a7/barbican-keystone-listener/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.050500 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5b8ccbfb5d-khqh7_d6806fde-b186-4daa-b56d-e746704bf9a7/barbican-keystone-listener-log/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.128849 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bb59858c-hh8zx_0126d82a-0297-4a4e-b75a-1e0a8f31d9d4/barbican-worker/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.176202 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-77bb59858c-hh8zx_0126d82a-0297-4a4e-b75a-1e0a8f31d9d4/barbican-worker-log/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.373156 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-s95wx_441768e0-9ee7-4d91-a9e1-d72ef8fb47e3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.419233 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/ceilometer-central-agent/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.495522 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/ceilometer-notification-agent/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.552621 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/proxy-httpd/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.589381 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_d6a77283-bbb4-4a6b-9713-f569e57ed673/sg-core/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.728411 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10903b59-1315-45e1-b734-40c85cfa17e7/cinder-api/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.773378 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_10903b59-1315-45e1-b734-40c85cfa17e7/cinder-api-log/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.930539 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b221b0fe-0514-409e-9ee6-966c6ceeb683/cinder-scheduler/0.log" Oct 09 09:21:47 crc kubenswrapper[4872]: I1009 09:21:47.951834 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_b221b0fe-0514-409e-9ee6-966c6ceeb683/probe/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.092804 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-9kpvg_57a6c64a-020d-4b90-b350-84f9441d08b3/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.171701 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-f24mq_f0febe64-b028-4782-a3e8-31b07031c230/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.283403 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-lr4lv_cc650d56-3744-41ca-90b1-707c8fc60bd9/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.366314 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/init/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.561251 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/init/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.596966 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-8c6f6df99-bd65m_1d643d17-4e67-41f2-a8fc-f9790033d364/dnsmasq-dns/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.651206 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-lrqqt_9b5aa1fc-253f-4c3b-b3f4-1364130a3a08/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.793070 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d651cc86-02b8-4d90-b4a0-3d1a2de1e300/glance-httpd/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.831871 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_d651cc86-02b8-4d90-b4a0-3d1a2de1e300/glance-log/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.970520 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a70f38a0-6bd6-45c9-8dc6-6f540842a4be/glance-log/0.log" Oct 09 09:21:48 crc kubenswrapper[4872]: I1009 09:21:48.981084 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_a70f38a0-6bd6-45c9-8dc6-6f540842a4be/glance-httpd/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.145408 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bf66bd496-2sdwc_6d37cb10-16dd-4360-aaf9-b90839973688/horizon/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.303507 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-xmr4n_abc08f25-2678-4241-9f86-d7c6700a7ae6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.477141 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-qdjb4_eb153425-cc47-4785-be11-49785db1cca7/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.563274 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5bf66bd496-2sdwc_6d37cb10-16dd-4360-aaf9-b90839973688/horizon-log/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.763021 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7489b48876-4zz4g_7f746daf-5432-4790-961e-c02910ec5946/keystone-api/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.823631 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29333341-v6cn6_3c82d19d-ec17-47d9-aace-a14cfb593965/keystone-cron/0.log" Oct 09 09:21:49 crc kubenswrapper[4872]: I1009 09:21:49.912149 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_4d540b3a-5230-484c-b807-ce072ce235d9/kube-state-metrics/0.log" Oct 09 09:21:50 crc kubenswrapper[4872]: I1009 09:21:50.018910 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-dvntq_80ca0a43-215b-42e4-8278-c7ded62e8080/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:50 crc kubenswrapper[4872]: I1009 09:21:50.407937 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c68d5ff89-tcgwk_7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0/neutron-api/0.log" Oct 09 09:21:50 crc kubenswrapper[4872]: I1009 09:21:50.435819 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-c68d5ff89-tcgwk_7ec7531f-ab27-4417-a1f3-8b9a8f9fd7f0/neutron-httpd/0.log" Oct 09 09:21:50 crc kubenswrapper[4872]: I1009 09:21:50.545498 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nb6pl_92e98033-47ad-4e23-a3c4-d8c886be6d0c/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.106946 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7a95ce78-4b3a-4aea-8351-c39dc3836f4c/nova-api-log/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.229564 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_c572e5bd-51c3-49b9-9c55-7e9f888926a2/nova-cell0-conductor-conductor/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.509577 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_cf29706b-8cae-42b7-ba76-b52189349378/nova-cell1-conductor-conductor/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.513916 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_7a95ce78-4b3a-4aea-8351-c39dc3836f4c/nova-api-api/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.558355 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_15c4d286-8c8e-40a2-b034-77f11abf9a08/nova-cell1-novncproxy-novncproxy/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.826474 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-h2wx7_fc88437f-6937-47eb-bebd-514527660dd2/nova-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:51 crc kubenswrapper[4872]: I1009 09:21:51.859555 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6f6917d5-6f1e-4f7f-b651-14ee02e1582c/nova-metadata-log/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.242745 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/mysql-bootstrap/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.256474 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_605ee569-e22f-4c02-8a31-34a9a92d7497/nova-scheduler-scheduler/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.420497 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/mysql-bootstrap/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.533272 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_373d202a-3ea1-4855-a4d5-938451b3c42d/galera/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.690711 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/mysql-bootstrap/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.828328 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/mysql-bootstrap/0.log" Oct 09 09:21:52 crc kubenswrapper[4872]: I1009 09:21:52.872881 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_0932ae43-84f6-4a0f-bf6e-fccd2f31c236/galera/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.012892 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_49ba1d52-eb34-49dd-8e46-34219f2e8336/openstackclient/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.163910 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_6f6917d5-6f1e-4f7f-b651-14ee02e1582c/nova-metadata-metadata/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.224934 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-7p7x8_19bf463a-82b1-4522-a8c7-0661281c1e50/openstack-network-exporter/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.348637 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server-init/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.553256 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server-init/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.563381 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovs-vswitchd/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.584059 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mfxrr_2a0f279e-d3c4-4d1a-b519-996c26f5ffba/ovsdb-server/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.815094 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-zk2nb_2e454e39-b253-4330-b5f7-7619c2f0f1c9/ovn-controller/0.log" Oct 09 09:21:53 crc kubenswrapper[4872]: I1009 09:21:53.820725 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-wwvpw_d9ac9cf3-10a8-45ef-b729-4822000bdc63/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.008115 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85/openstack-network-exporter/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.008935 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_4f7b58e0-4eea-4ec2-b3e3-df0ccdf9aa85/ovn-northd/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.187112 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a/openstack-network-exporter/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.244209 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_f2fbc29a-eb3d-46bc-9bd8-10f4ce7b3c5a/ovsdbserver-nb/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.353526 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_81338d5d-c3ed-4163-90c3-0e60e578e35b/openstack-network-exporter/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.450534 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_81338d5d-c3ed-4163-90c3-0e60e578e35b/ovsdbserver-sb/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.702518 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66db9d8d8d-cz7s2_f4ad5d2f-42db-4d64-9d37-d37e89539c20/placement-api/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.727216 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/setup-container/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.757813 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-66db9d8d8d-cz7s2_f4ad5d2f-42db-4d64-9d37-d37e89539c20/placement-log/0.log" Oct 09 09:21:54 crc kubenswrapper[4872]: I1009 09:21:54.962814 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/rabbitmq/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.026467 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_06d9f9e2-b9c5-4a08-b88c-c94e66a09fcc/setup-container/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.068703 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/setup-container/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.188200 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/setup-container/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.278382 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_b180311b-aa96-4a46-bbb2-5c32d18b0593/rabbitmq/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.295132 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-7dkll_710632e7-e59c-4437-a9c5-171a80e989f6/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.520267 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qh4f6_b21bf250-3cfe-47da-8ff1-2c4a54dc7b77/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.538388 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-qd5jg_5bf8bedf-9b23-4570-b728-fc01c39c9373/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.703780 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-w99z2_0a5b2694-c0aa-44f4-aec1-4b3dad1a750d/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:55 crc kubenswrapper[4872]: I1009 09:21:55.782442 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-h292d_cefe7874-2185-4237-b9b6-aa8332849f11/ssh-known-hosts-edpm-deployment/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.013820 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b44c8599-9mlwc_dbc66806-aef2-4aa8-86ca-c48f90386f6c/proxy-server/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.069662 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-5b44c8599-9mlwc_dbc66806-aef2-4aa8-86ca-c48f90386f6c/proxy-httpd/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.134404 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-qx7q8_31fd23b6-fa18-4681-ba21-b9e69ad958d0/swift-ring-rebalance/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.240128 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-auditor/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.300496 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-reaper/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.393395 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-replicator/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.462002 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:21:56 crc kubenswrapper[4872]: E1009 09:21:56.462247 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.495378 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/account-server/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.504193 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-auditor/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.527319 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-replicator/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.621809 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-server/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.690223 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-auditor/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.736883 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-expirer/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.744006 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/container-updater/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.886326 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-replicator/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.887576 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-server/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.957293 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/object-updater/0.log" Oct 09 09:21:56 crc kubenswrapper[4872]: I1009 09:21:56.969558 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/rsync/0.log" Oct 09 09:21:57 crc kubenswrapper[4872]: I1009 09:21:57.092000 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_6ee9bdf9-4380-47b9-b1b9-083974d8f01a/swift-recon-cron/0.log" Oct 09 09:21:57 crc kubenswrapper[4872]: I1009 09:21:57.227823 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-258q8_55becb4c-820e-4d84-b7c3-e3b2bddc84a0/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:21:57 crc kubenswrapper[4872]: I1009 09:21:57.340273 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_efed60a8-77a0-4635-8ea2-46f3400b9090/tempest-tests-tempest-tests-runner/0.log" Oct 09 09:21:57 crc kubenswrapper[4872]: I1009 09:21:57.419022 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_cd9e08b7-6f45-42d6-be93-5534318c54cd/test-operator-logs-container/0.log" Oct 09 09:21:57 crc kubenswrapper[4872]: I1009 09:21:57.561305 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-msgrc_3586ab35-c3cc-4f05-8319-b0b568c2c029/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Oct 09 09:22:06 crc kubenswrapper[4872]: I1009 09:22:06.347224 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_5c8202d4-bb37-4cad-a36c-547b55adf925/memcached/0.log" Oct 09 09:22:10 crc kubenswrapper[4872]: I1009 09:22:10.467485 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:22:10 crc kubenswrapper[4872]: E1009 09:22:10.469145 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.509981 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.749907 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.760911 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.768290 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.966993 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/pull/0.log" Oct 09 09:22:20 crc kubenswrapper[4872]: I1009 09:22:20.990566 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/extract/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.012862 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_01a6fbbbfebe13815ff5c4cf9715d74736160b159161e163d0c6a62f78gg8bs_6e67af22-d4fc-40de-b757-d08690f67c9c/util/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.399142 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-lrlzc_91db56d0-f7bc-4d02-92ec-c7ead6625176/manager/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.453039 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-vsdt6_283f56cf-99e9-471b-8f40-10ec0f31ad06/kube-rbac-proxy/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.456145 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-lrlzc_91db56d0-f7bc-4d02-92ec-c7ead6625176/kube-rbac-proxy/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.461392 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:22:21 crc kubenswrapper[4872]: E1009 09:22:21.461696 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.599102 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-vsdt6_283f56cf-99e9-471b-8f40-10ec0f31ad06/manager/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.635478 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-26dhq_0d04d185-02c6-4688-83cd-e9f1744776f1/kube-rbac-proxy/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.662151 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-26dhq_0d04d185-02c6-4688-83cd-e9f1744776f1/manager/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.815350 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-dskmk_fab5d9a4-352a-4092-9139-c74242142eda/kube-rbac-proxy/0.log" Oct 09 09:22:21 crc kubenswrapper[4872]: I1009 09:22:21.907877 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-dskmk_fab5d9a4-352a-4092-9139-c74242142eda/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.018024 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-qnnp4_b9b5d69d-fb26-4286-a9b0-3b00d042972a/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.024571 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-qnnp4_b9b5d69d-fb26-4286-a9b0-3b00d042972a/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.164984 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-n5rbg_deb85a9b-515a-4027-a178-c83829fd5a34/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.291382 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-n5rbg_deb85a9b-515a-4027-a178-c83829fd5a34/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.366941 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-kgzkf_57a99d6c-df10-4de9-9434-471ceb56878d/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.515285 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-585fc5b659-kgzkf_57a99d6c-df10-4de9-9434-471ceb56878d/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.524796 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vqj98_c5d3f64a-0214-457c-b4c6-f48beabe8669/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.625306 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-74cb5cbc49-vqj98_c5d3f64a-0214-457c-b4c6-f48beabe8669/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.730978 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-8rkjv_a34a57a1-2b91-4d65-8336-4a363b33dd45/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.771738 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-8rkjv_a34a57a1-2b91-4d65-8336-4a363b33dd45/manager/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.901789 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-2knwl_18e70210-0a78-4ea0-8d8b-a955cdbec6b8/kube-rbac-proxy/0.log" Oct 09 09:22:22 crc kubenswrapper[4872]: I1009 09:22:22.936322 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-2knwl_18e70210-0a78-4ea0-8d8b-a955cdbec6b8/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.031387 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-8ttcw_88a9a65b-d7fe-4471-b408-562f5c98bdbb/kube-rbac-proxy/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.089717 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-8ttcw_88a9a65b-d7fe-4471-b408-562f5c98bdbb/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.190046 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-kz58p_fcb8eabd-193f-4b03-b011-5e18514cc858/kube-rbac-proxy/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.280781 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-kz58p_fcb8eabd-193f-4b03-b011-5e18514cc858/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.401395 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-v67sz_ae4070fd-f49f-403d-a0b5-3285f235a221/kube-rbac-proxy/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.489056 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-v67sz_ae4070fd-f49f-403d-a0b5-3285f235a221/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.553365 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-ffks5_c4fa74e8-0166-4704-8a06-401b53e1cefe/kube-rbac-proxy/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.626614 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-ffks5_c4fa74e8-0166-4704-8a06-401b53e1cefe/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.690516 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc_4099adc7-9a49-4191-bfc3-3f17b15312db/kube-rbac-proxy/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.750472 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dm66dc_4099adc7-9a49-4191-bfc3-3f17b15312db/manager/0.log" Oct 09 09:22:23 crc kubenswrapper[4872]: I1009 09:22:23.894174 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f665bd66d-9lhnp_4dfe7dfe-548f-4ec8-a435-72217082dd3b/kube-rbac-proxy/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.095030 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-557bb754c5-pj5t9_02685dc1-985a-4360-8192-74858bb6062b/kube-rbac-proxy/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.268825 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-557bb754c5-pj5t9_02685dc1-985a-4360-8192-74858bb6062b/operator/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.311863 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-5jngg_6249ff92-0786-431e-9f18-d48d3b2a6863/registry-server/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.471094 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-xphlj_6c78484d-5fa5-4f18-ace9-a6811363dca7/kube-rbac-proxy/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.532880 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-6f96f8c84-xphlj_6c78484d-5fa5-4f18-ace9-a6811363dca7/manager/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.657802 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-sbzmh_25798330-695b-4a9d-b271-48b6d025a3f8/kube-rbac-proxy/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.741700 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-sbzmh_25798330-695b-4a9d-b271-48b6d025a3f8/manager/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.898556 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-qwhbz_add3400b-1fe2-4b5e-a810-7cca5a38ac96/operator/0.log" Oct 09 09:22:24 crc kubenswrapper[4872]: I1009 09:22:24.947533 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5f665bd66d-9lhnp_4dfe7dfe-548f-4ec8-a435-72217082dd3b/manager/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.025387 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-w6jlm_bfaf85e3-55b3-4564-9ffc-3b80e7f32c85/manager/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.065839 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-w6jlm_bfaf85e3-55b3-4564-9ffc-3b80e7f32c85/kube-rbac-proxy/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.206248 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-b9zg2_a9091168-5676-4cd3-9a0d-f76d0c7c88df/kube-rbac-proxy/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.245074 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-b9zg2_a9091168-5676-4cd3-9a0d-f76d0c7c88df/manager/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.328196 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-ppxzv_0cf7b66d-2680-43e9-812b-1a3713eea708/kube-rbac-proxy/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.351145 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-ppxzv_0cf7b66d-2680-43e9-812b-1a3713eea708/manager/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.479701 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjx2l_3b929d55-693e-419d-ad6e-aa0005f6538a/manager/0.log" Oct 09 09:22:25 crc kubenswrapper[4872]: I1009 09:22:25.483147 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-sjx2l_3b929d55-693e-419d-ad6e-aa0005f6538a/kube-rbac-proxy/0.log" Oct 09 09:22:26 crc kubenswrapper[4872]: I1009 09:22:26.976866 4872 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:26 crc kubenswrapper[4872]: E1009 09:22:26.977369 4872 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0cac16d-6629-4d42-a83b-305bd1356f48" containerName="container-00" Oct 09 09:22:26 crc kubenswrapper[4872]: I1009 09:22:26.977389 4872 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0cac16d-6629-4d42-a83b-305bd1356f48" containerName="container-00" Oct 09 09:22:26 crc kubenswrapper[4872]: I1009 09:22:26.977699 4872 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0cac16d-6629-4d42-a83b-305bd1356f48" containerName="container-00" Oct 09 09:22:26 crc kubenswrapper[4872]: I1009 09:22:26.979387 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:26 crc kubenswrapper[4872]: I1009 09:22:26.990609 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.053585 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrc4k\" (UniqueName: \"kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.053677 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.053812 4872 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.155331 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrc4k\" (UniqueName: \"kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.155414 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.155559 4872 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.156546 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.156706 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.187987 4872 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrc4k\" (UniqueName: \"kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k\") pod \"redhat-marketplace-fjlhf\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.300782 4872 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:27 crc kubenswrapper[4872]: I1009 09:22:27.773177 4872 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:28 crc kubenswrapper[4872]: I1009 09:22:28.324427 4872 generic.go:334] "Generic (PLEG): container finished" podID="49d59d7b-ee2e-47f4-8134-d9450bd147e0" containerID="3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909" exitCode=0 Oct 09 09:22:28 crc kubenswrapper[4872]: I1009 09:22:28.324499 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerDied","Data":"3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909"} Oct 09 09:22:28 crc kubenswrapper[4872]: I1009 09:22:28.326171 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerStarted","Data":"0d277ea8eb129ed1a250de8b276e0dc6ef8ff2ac59e1a1705a4749bebe876889"} Oct 09 09:22:29 crc kubenswrapper[4872]: I1009 09:22:29.340844 4872 generic.go:334] "Generic (PLEG): container finished" podID="49d59d7b-ee2e-47f4-8134-d9450bd147e0" containerID="ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7" exitCode=0 Oct 09 09:22:29 crc kubenswrapper[4872]: I1009 09:22:29.341418 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerDied","Data":"ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7"} Oct 09 09:22:30 crc kubenswrapper[4872]: I1009 09:22:30.352373 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerStarted","Data":"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333"} Oct 09 09:22:30 crc kubenswrapper[4872]: I1009 09:22:30.371349 4872 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fjlhf" podStartSLOduration=2.694595873 podStartE2EDuration="4.371327394s" podCreationTimestamp="2025-10-09 09:22:26 +0000 UTC" firstStartedPulling="2025-10-09 09:22:28.325860917 +0000 UTC m=+3906.516389543" lastFinishedPulling="2025-10-09 09:22:30.002592438 +0000 UTC m=+3908.193121064" observedRunningTime="2025-10-09 09:22:30.367712838 +0000 UTC m=+3908.558241464" watchObservedRunningTime="2025-10-09 09:22:30.371327394 +0000 UTC m=+3908.561856020" Oct 09 09:22:36 crc kubenswrapper[4872]: I1009 09:22:36.461964 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:22:36 crc kubenswrapper[4872]: E1009 09:22:36.462901 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:22:37 crc kubenswrapper[4872]: I1009 09:22:37.301795 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:37 crc kubenswrapper[4872]: I1009 09:22:37.302177 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:37 crc kubenswrapper[4872]: I1009 09:22:37.351200 4872 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:37 crc kubenswrapper[4872]: I1009 09:22:37.455880 4872 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:37 crc kubenswrapper[4872]: I1009 09:22:37.581816 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:39 crc kubenswrapper[4872]: I1009 09:22:39.427325 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fjlhf" podUID="49d59d7b-ee2e-47f4-8134-d9450bd147e0" containerName="registry-server" containerID="cri-o://cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333" gracePeriod=2 Oct 09 09:22:39 crc kubenswrapper[4872]: I1009 09:22:39.907959 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:39 crc kubenswrapper[4872]: I1009 09:22:39.990101 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrc4k\" (UniqueName: \"kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k\") pod \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " Oct 09 09:22:39 crc kubenswrapper[4872]: I1009 09:22:39.990302 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities\") pod \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " Oct 09 09:22:39 crc kubenswrapper[4872]: I1009 09:22:39.990403 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content\") pod \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\" (UID: \"49d59d7b-ee2e-47f4-8134-d9450bd147e0\") " Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.002025 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities" (OuterVolumeSpecName: "utilities") pod "49d59d7b-ee2e-47f4-8134-d9450bd147e0" (UID: "49d59d7b-ee2e-47f4-8134-d9450bd147e0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.021962 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k" (OuterVolumeSpecName: "kube-api-access-qrc4k") pod "49d59d7b-ee2e-47f4-8134-d9450bd147e0" (UID: "49d59d7b-ee2e-47f4-8134-d9450bd147e0"). InnerVolumeSpecName "kube-api-access-qrc4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.024385 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "49d59d7b-ee2e-47f4-8134-d9450bd147e0" (UID: "49d59d7b-ee2e-47f4-8134-d9450bd147e0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.092585 4872 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-utilities\") on node \"crc\" DevicePath \"\"" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.092633 4872 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/49d59d7b-ee2e-47f4-8134-d9450bd147e0-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.092663 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrc4k\" (UniqueName: \"kubernetes.io/projected/49d59d7b-ee2e-47f4-8134-d9450bd147e0-kube-api-access-qrc4k\") on node \"crc\" DevicePath \"\"" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.438884 4872 generic.go:334] "Generic (PLEG): container finished" podID="49d59d7b-ee2e-47f4-8134-d9450bd147e0" containerID="cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333" exitCode=0 Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.438926 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerDied","Data":"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333"} Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.438951 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjlhf" event={"ID":"49d59d7b-ee2e-47f4-8134-d9450bd147e0","Type":"ContainerDied","Data":"0d277ea8eb129ed1a250de8b276e0dc6ef8ff2ac59e1a1705a4749bebe876889"} Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.438953 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjlhf" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.438968 4872 scope.go:117] "RemoveContainer" containerID="cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.468924 4872 scope.go:117] "RemoveContainer" containerID="ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.496908 4872 scope.go:117] "RemoveContainer" containerID="3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.503538 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.520070 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjlhf"] Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.553900 4872 scope.go:117] "RemoveContainer" containerID="cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333" Oct 09 09:22:40 crc kubenswrapper[4872]: E1009 09:22:40.554422 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333\": container with ID starting with cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333 not found: ID does not exist" containerID="cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.554475 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333"} err="failed to get container status \"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333\": rpc error: code = NotFound desc = could not find container \"cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333\": container with ID starting with cd947392ae34178e31d86b31b522bea654f15c7d68ad786e243ac542e74a3333 not found: ID does not exist" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.554506 4872 scope.go:117] "RemoveContainer" containerID="ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7" Oct 09 09:22:40 crc kubenswrapper[4872]: E1009 09:22:40.554969 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7\": container with ID starting with ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7 not found: ID does not exist" containerID="ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.554995 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7"} err="failed to get container status \"ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7\": rpc error: code = NotFound desc = could not find container \"ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7\": container with ID starting with ae65b1ba46906ac23c1471d3f2bcf1efd9ead49d902c30295312d4adab157dd7 not found: ID does not exist" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.555009 4872 scope.go:117] "RemoveContainer" containerID="3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909" Oct 09 09:22:40 crc kubenswrapper[4872]: E1009 09:22:40.555202 4872 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909\": container with ID starting with 3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909 not found: ID does not exist" containerID="3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909" Oct 09 09:22:40 crc kubenswrapper[4872]: I1009 09:22:40.555228 4872 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909"} err="failed to get container status \"3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909\": rpc error: code = NotFound desc = could not find container \"3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909\": container with ID starting with 3132323a79f4d0f02dd0476588fb0f9cece809354ffc3d3e8f07cf82af494909 not found: ID does not exist" Oct 09 09:22:41 crc kubenswrapper[4872]: I1009 09:22:41.615970 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-b5qvz_1f314406-7129-4061-922a-7c2fba4019d9/control-plane-machine-set-operator/0.log" Oct 09 09:22:41 crc kubenswrapper[4872]: I1009 09:22:41.767141 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wcb54_a1a5b7f5-7caa-4fe2-afb4-6c7176939e63/kube-rbac-proxy/0.log" Oct 09 09:22:41 crc kubenswrapper[4872]: I1009 09:22:41.772975 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-wcb54_a1a5b7f5-7caa-4fe2-afb4-6c7176939e63/machine-api-operator/0.log" Oct 09 09:22:42 crc kubenswrapper[4872]: I1009 09:22:42.487025 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49d59d7b-ee2e-47f4-8134-d9450bd147e0" path="/var/lib/kubelet/pods/49d59d7b-ee2e-47f4-8134-d9450bd147e0/volumes" Oct 09 09:22:49 crc kubenswrapper[4872]: I1009 09:22:49.461817 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:22:49 crc kubenswrapper[4872]: E1009 09:22:49.462671 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:22:52 crc kubenswrapper[4872]: I1009 09:22:52.704845 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-cq5wd_ba374ef8-6f28-4a47-a948-8cd66cff6132/cert-manager-controller/0.log" Oct 09 09:22:52 crc kubenswrapper[4872]: I1009 09:22:52.821297 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r594q_eb652e7d-6150-46ae-956f-0fcf642d935a/cert-manager-cainjector/0.log" Oct 09 09:22:52 crc kubenswrapper[4872]: I1009 09:22:52.908251 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-k456p_f1f5f354-df1e-49fa-a117-8e99befeaf38/cert-manager-webhook/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.373659 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-fz8l7_d9654e2a-2300-47ca-8725-682e7d1cee0a/nmstate-console-plugin/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.517821 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-lgwck_591e63c4-2cfa-478c-a509-e6ad87de2cb5/nmstate-handler/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.550631 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fn24v_40f6895e-a74d-4ef0-bb3b-9024d4b2db5d/kube-rbac-proxy/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.589204 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-fn24v_40f6895e-a74d-4ef0-bb3b-9024d4b2db5d/nmstate-metrics/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.731595 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-nwsgr_2d597183-b39b-40c1-ac08-ad5467522fc6/nmstate-operator/0.log" Oct 09 09:23:03 crc kubenswrapper[4872]: I1009 09:23:03.799053 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-ppr2b_d4b82a24-7a85-4e6b-9376-dc4dc2d72e2b/nmstate-webhook/0.log" Oct 09 09:23:04 crc kubenswrapper[4872]: I1009 09:23:04.462099 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:23:04 crc kubenswrapper[4872]: E1009 09:23:04.462440 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:23:17 crc kubenswrapper[4872]: I1009 09:23:17.461572 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:23:17 crc kubenswrapper[4872]: E1009 09:23:17.462424 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:23:18 crc kubenswrapper[4872]: I1009 09:23:18.844064 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2hm62_fadbcba0-ee7c-4e6b-8134-b60655294553/kube-rbac-proxy/0.log" Oct 09 09:23:18 crc kubenswrapper[4872]: I1009 09:23:18.913078 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-2hm62_fadbcba0-ee7c-4e6b-8134-b60655294553/controller/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.015201 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.204822 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.206561 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.223787 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.245776 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.915797 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.953704 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:23:19 crc kubenswrapper[4872]: I1009 09:23:19.956127 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.012731 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.153398 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-frr-files/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.200605 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-reloader/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.218465 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/cp-metrics/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.237827 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/controller/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.402607 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/kube-rbac-proxy/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.425122 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/frr-metrics/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.479008 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/kube-rbac-proxy-frr/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.664745 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/reloader/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.734145 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-9zqhg_381ccfd5-05cf-4592-933e-128d714bf0a6/frr-k8s-webhook-server/0.log" Oct 09 09:23:20 crc kubenswrapper[4872]: I1009 09:23:20.978204 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d774594fc-m77qs_75ddc562-e0fb-4c1c-b98d-eb754c97b989/manager/0.log" Oct 09 09:23:21 crc kubenswrapper[4872]: I1009 09:23:21.200454 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f4fb54f85-5jlsq_c49185d6-a699-454b-8e58-88d86a55b694/webhook-server/0.log" Oct 09 09:23:21 crc kubenswrapper[4872]: I1009 09:23:21.236471 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-b52wr_76fe700a-cada-4802-91e5-1bd3522ab565/kube-rbac-proxy/0.log" Oct 09 09:23:21 crc kubenswrapper[4872]: I1009 09:23:21.844381 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hd8n4_94c5800f-97ae-4482-a959-d459cbdfdbaf/frr/0.log" Oct 09 09:23:21 crc kubenswrapper[4872]: I1009 09:23:21.850377 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-b52wr_76fe700a-cada-4802-91e5-1bd3522ab565/speaker/0.log" Oct 09 09:23:30 crc kubenswrapper[4872]: I1009 09:23:30.462149 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:23:30 crc kubenswrapper[4872]: E1009 09:23:30.463064 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:23:32 crc kubenswrapper[4872]: I1009 09:23:32.685481 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:23:32 crc kubenswrapper[4872]: I1009 09:23:32.820064 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:23:32 crc kubenswrapper[4872]: I1009 09:23:32.835266 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:23:32 crc kubenswrapper[4872]: I1009 09:23:32.865747 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.033741 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/pull/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.050737 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/util/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.076105 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2xphwm_fbdc30ed-ca91-4dc0-b908-a6781a0857e4/extract/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.234317 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.361351 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.374649 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.402913 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.539899 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-utilities/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.583767 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/extract-content/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.750897 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:23:33 crc kubenswrapper[4872]: I1009 09:23:33.928591 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-tj85s_ee093ff4-27c6-46de-abb9-79df201ca9ea/registry-server/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.002329 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.028109 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.035661 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.186780 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-content/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.256389 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/extract-utilities/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.388565 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.626983 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.668807 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.698979 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.699749 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-x8nzf_f41d6540-7af0-4aa7-a36a-629e55c85a6f/registry-server/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.841991 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/util/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.863064 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/extract/0.log" Oct 09 09:23:34 crc kubenswrapper[4872]: I1009 09:23:34.883278 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835c7nlrn_d492de09-5100-4997-b375-2e60812f4591/pull/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.022676 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-69jm8_6b2d20f2-e34f-4b19-85fb-0e0bfffe849d/marketplace-operator/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.084991 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.281960 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.322718 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.339842 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.459889 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-utilities/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.541585 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/extract-content/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.626767 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-8swgj_67dc2250-a671-4f10-8ac9-46021dca556b/registry-server/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.668170 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.827497 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.852334 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:23:35 crc kubenswrapper[4872]: I1009 09:23:35.865376 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:23:36 crc kubenswrapper[4872]: I1009 09:23:36.029526 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-utilities/0.log" Oct 09 09:23:36 crc kubenswrapper[4872]: I1009 09:23:36.095017 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/extract-content/0.log" Oct 09 09:23:36 crc kubenswrapper[4872]: I1009 09:23:36.434480 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mdw5n_c9a7cca0-9eab-4de5-8100-98cf3f887af1/registry-server/0.log" Oct 09 09:23:45 crc kubenswrapper[4872]: I1009 09:23:45.462272 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:23:45 crc kubenswrapper[4872]: E1009 09:23:45.463112 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:23:57 crc kubenswrapper[4872]: I1009 09:23:57.461979 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:23:57 crc kubenswrapper[4872]: E1009 09:23:57.462921 4872 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-98kmz_openshift-machine-config-operator(5cc7b6a1-93b1-455b-aa21-d8bb7574857e)\"" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" Oct 09 09:24:09 crc kubenswrapper[4872]: I1009 09:24:09.461479 4872 scope.go:117] "RemoveContainer" containerID="252d786528562f6c99a9957e4eeefb3e2413bae7257615a6d60d02cae8db0a94" Oct 09 09:24:10 crc kubenswrapper[4872]: I1009 09:24:10.293624 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" event={"ID":"5cc7b6a1-93b1-455b-aa21-d8bb7574857e","Type":"ContainerStarted","Data":"489951eb123dec8cb4cc85c0710d60351af4dd3907a395500a7133e54feef50b"} Oct 09 09:25:12 crc kubenswrapper[4872]: I1009 09:25:12.968083 4872 generic.go:334] "Generic (PLEG): container finished" podID="c8f8a617-7e86-49df-be7f-ccf469601c9a" containerID="49d405d791525b46078eefea6e5dc82a657ff9a87abdc7f9066bed4de05542b2" exitCode=0 Oct 09 09:25:12 crc kubenswrapper[4872]: I1009 09:25:12.968189 4872 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-84zmz/must-gather-cdxhx" event={"ID":"c8f8a617-7e86-49df-be7f-ccf469601c9a","Type":"ContainerDied","Data":"49d405d791525b46078eefea6e5dc82a657ff9a87abdc7f9066bed4de05542b2"} Oct 09 09:25:12 crc kubenswrapper[4872]: I1009 09:25:12.969282 4872 scope.go:117] "RemoveContainer" containerID="49d405d791525b46078eefea6e5dc82a657ff9a87abdc7f9066bed4de05542b2" Oct 09 09:25:13 crc kubenswrapper[4872]: I1009 09:25:13.937912 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-84zmz_must-gather-cdxhx_c8f8a617-7e86-49df-be7f-ccf469601c9a/gather/0.log" Oct 09 09:25:23 crc kubenswrapper[4872]: I1009 09:25:23.931675 4872 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-84zmz/must-gather-cdxhx"] Oct 09 09:25:23 crc kubenswrapper[4872]: I1009 09:25:23.932752 4872 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-84zmz/must-gather-cdxhx" podUID="c8f8a617-7e86-49df-be7f-ccf469601c9a" containerName="copy" containerID="cri-o://2df4b4d4e29304492834013ec6e632691ccfca48b85a131ad22e33ea22cf91d8" gracePeriod=2 Oct 09 09:25:23 crc kubenswrapper[4872]: I1009 09:25:23.939313 4872 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-84zmz/must-gather-cdxhx"] Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.092844 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-84zmz_must-gather-cdxhx_c8f8a617-7e86-49df-be7f-ccf469601c9a/copy/0.log" Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.093321 4872 generic.go:334] "Generic (PLEG): container finished" podID="c8f8a617-7e86-49df-be7f-ccf469601c9a" containerID="2df4b4d4e29304492834013ec6e632691ccfca48b85a131ad22e33ea22cf91d8" exitCode=143 Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.735441 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-84zmz_must-gather-cdxhx_c8f8a617-7e86-49df-be7f-ccf469601c9a/copy/0.log" Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.736219 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.913509 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v52xv\" (UniqueName: \"kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv\") pod \"c8f8a617-7e86-49df-be7f-ccf469601c9a\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.913997 4872 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output\") pod \"c8f8a617-7e86-49df-be7f-ccf469601c9a\" (UID: \"c8f8a617-7e86-49df-be7f-ccf469601c9a\") " Oct 09 09:25:24 crc kubenswrapper[4872]: I1009 09:25:24.981976 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv" (OuterVolumeSpecName: "kube-api-access-v52xv") pod "c8f8a617-7e86-49df-be7f-ccf469601c9a" (UID: "c8f8a617-7e86-49df-be7f-ccf469601c9a"). InnerVolumeSpecName "kube-api-access-v52xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.019337 4872 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v52xv\" (UniqueName: \"kubernetes.io/projected/c8f8a617-7e86-49df-be7f-ccf469601c9a-kube-api-access-v52xv\") on node \"crc\" DevicePath \"\"" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.108940 4872 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-84zmz_must-gather-cdxhx_c8f8a617-7e86-49df-be7f-ccf469601c9a/copy/0.log" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.109557 4872 scope.go:117] "RemoveContainer" containerID="2df4b4d4e29304492834013ec6e632691ccfca48b85a131ad22e33ea22cf91d8" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.109582 4872 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-84zmz/must-gather-cdxhx" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.111153 4872 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "c8f8a617-7e86-49df-be7f-ccf469601c9a" (UID: "c8f8a617-7e86-49df-be7f-ccf469601c9a"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.121706 4872 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/c8f8a617-7e86-49df-be7f-ccf469601c9a-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 09 09:25:25 crc kubenswrapper[4872]: I1009 09:25:25.130786 4872 scope.go:117] "RemoveContainer" containerID="49d405d791525b46078eefea6e5dc82a657ff9a87abdc7f9066bed4de05542b2" Oct 09 09:25:26 crc kubenswrapper[4872]: I1009 09:25:26.472777 4872 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8f8a617-7e86-49df-be7f-ccf469601c9a" path="/var/lib/kubelet/pods/c8f8a617-7e86-49df-be7f-ccf469601c9a/volumes" Oct 09 09:26:38 crc kubenswrapper[4872]: I1009 09:26:38.156596 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:26:38 crc kubenswrapper[4872]: I1009 09:26:38.157259 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 09 09:27:08 crc kubenswrapper[4872]: I1009 09:27:08.157238 4872 patch_prober.go:28] interesting pod/machine-config-daemon-98kmz container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 09 09:27:08 crc kubenswrapper[4872]: I1009 09:27:08.157804 4872 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-98kmz" podUID="5cc7b6a1-93b1-455b-aa21-d8bb7574857e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515071677773024471 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015071677774017407 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015071667307016521 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015071667307015471 5ustar corecore